var/home/core/zuul-output/0000755000175000017500000000000015145546172014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015145557015015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000324516315145556677020311 0ustar corecoreݖikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB ?YI_翪|mvşo#oVݏKf+ovpZj3C4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;FmZl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) *ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&c(޸0"$5ڪҾη*t:%?vEmO5tqÜ3Cyu '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?}=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@4ޙ-did˥]5]5᪩QJlyIPEQZȰ<'$VO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*%WyΈ W\Of[=߰+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^sнM\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bd+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTtx袥:2JޚݶKd7UZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOa9C1L ={fm&'^tigk$DA' elW@Tiv{ !]oBLKJO*t*\n-iȚ4`{x_z;j3Xh ׄ?xt.o:`x^d~0u$ v48 0_ | E"Hd"H`A0&dY3 ً[fctWF_hdxMUY.b=eaI3Z=᢬-'~DWc;j FRrI5%N/K;Dk rCbm7чsSW_8g{RY.~XfEߪg:smBi1 YBX4),[c^54Sg(s$sN' 88`wC3TE+A\.ԍל9 y{͝BxG&JS meT;{З>'[LR"w F05N<&AJ3DA0ʄ4(zTUWDdE3̻l^-Xw3Fɀ{B-~.h+U8 i1b8wؖ#~zQ`/L 9#Pu/<4A L<KL U(Ee'sCcq !Ȥ4΍ +aM(VldX ][T !Ȱ|HN~6y,⒊)$e{)SR#kהyϛ7^i58f4PmB8 Y{qeφvk73:1@ƛ.{f8IGv*1藺yx27M=>+VnG;\<x7v21՚H :[Γd!E'a4n?k[A׈(sob 41Y9(^SE@7`KIK`kx& V`X0,%pe_ן >hd xе"Q4SUwy x<'o_~#6$g!D$c=5ۄX[ു RzG:柺[ӏ[3frl ô ހ^2TӘUAT!94[[m۾\T)W> lv+ H\FpG)ۏjk_c51̃^cn ba-X/#=Im41NLu\9ETp^poAOO&Ack vz(vb$^Nyo$p[DtUCE9s".zɪ) ӓT)D:fci[*`cc&VhfFp佬)/Wdځ+ uR<$}Kr'ݔTW$md1"#mC_@:m P>DEu&ݛȘPˬ-Ő\B`xr`"F'Iٺ*DnA)yzr^!3Ír!S$,.:+d̋BʺJ#SX*8ҁW7~>oOFe-<uJQ|FZEP__gi(`0/ƍcv7go2G$ N%v$^^&Q 4AMbvvɀ1J{ڔhэK'9*W )IYO;E4z⛢79"hK{BFEmBAΛ3>IO j u߿d{=t-n3Pnef9[}=%G*9sX,¬xS&9'E&"/"ncx}"mV5tŘ:wcZ К G)]$mbXE ^ǽ8%>,0FЕ 6vAVKVCjrD25#Lrv?33Iam:xy`|Q'eű^\ơ' .gygSAixپ im41;P^azl5|JE2z=.wcMԧ ax& =`|#HQ*lS<.U׻`>ajϿ '!9MHK:9#s,jV剤C:LIeHJ"M8P,$N;a-zݸJWc :.<sR6 լ$gu4M*B(A ݖΑِ %H;S*ڳJt>$M!^*n3qESfU, Iĭb#UFJPvBgZvn aE5}~2E|=D' ܇q>8[¿yp/9Om/5|k \6xH.Z'OeCD@cq:Y~<1LٖY9# xe8g IKTQ:+Xg:*}.<M{ZH[^>m0G{ ̷hiOO|9Y"mma[sSbb'Rv&{@6; KE.a\}:<]Oyve3h9}E[kMD,5 %sO{킒 8.K?]i/`׎tp NvԻV4|<{H@#*h{Yp/E%dlh\bU:E%h@&SEK [ Ƣ xg{z%ǻViX~鮦w35QE~qp[ʕ@}ZL! Z0!A⼏q)[f &E1K3i+`JG P/EG 4 9LڑKL|`PОnG#|}qOR{Q|2_tH߫%pD?1%(@nfxOrs25rMլf{sk7݇fjӞh2HkeL'Wʿ}Ƞ%>9cSH|cEyQp 'ˢd:,v-us"Iidw>%zM@9IqrGq:&_p3õB!>9'0LL]M[lwWVR9I5YpVgtuZfG{RoZr3ٮr;wW:͋nqCRu1y=㊻Ij z[|W%q0 CJV٨3,ib{eH7 mҝ(3ɏO/̗-=OR\dIoHZ6n`R֑&#.Mv0vԬ]I˟vrK}F9X|FI#g.Gi)%!iK|o}|ֵ7!ېATJKB2Z/"BfB(gdj۸=}'),-iX'|M2roK\e5Pt:*qSH PgƉU'VKξ ,!3`˞tӱM'^Y&,mCM)eu㠥Ѻ\a}1:V1zMzT}R,IA e<%!vĉq|?mtB|A ?dXuWLGml?*uTC̶V`FVY>ECmDnG+UaKtȃbeb筃kݴO~f^⊈ 8MK?:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk56s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȯ.vIH"ŐR ;@~y>Kv{) 9AG ćͩ$.!б~N8i"1KФ\L7/,U@.ڮO?mُa ې!rGHw@56DǑq LA!&mYJ*ixz2*{_;IYJXFfQ* 0kA".mݡ"3`Rd1_u6d逖`7xGMf}k/⨼0Κ_pLq7k!dT x삖A7 u/~&ӄMu.<|yi I?@)XJ7{ޱ?Q]{#\4ZfR-dVaz./f+yGNMGOK?2_~3\z=y}^G$*A! IcuR.o=MZ9zu b#s9@*иrI@*qQN||Ix;I}&ݢ6ɢ}{]x}_o>Mm8S]~(EX{S/{&Ά+4"^'1lfYuB!!=8?8W:ZUa6-[N7 |Xpn1}nr }C5`J `rIJ;?08¢E WiBhF[|ݪSR3]J-=҅31,j2Y QH -]n_262;VI'ɮ}Lu>'$0&*m.)iHz|BvU0h} -W>7^ݔn,?WTm>C9|H nHNe"z`0A*0)QsUN8tl^N+mXU-q2EDöbVm')fpOj4r@mί%Y!/7$&5n8TK&cB/5q0){F74~'*{[\M-~#aSmMÉB2NnꝤʇ)b^g`u2x"8U [JYSk, "nu\h1Yhl~[mhm+F(g +YtHg ]7n]Q!Mǧę5bR!JbV>&w6οH+NL$]p>8Uu>Ѫ g39[=OF9V?SAT~:gGt $*}aQ.Zq~-K\rfm$%ɪ/%Wa}TUl|X- ,Bqn|&0o- ,Bqfrtx-rei0hE˝ݸDt#;I}8`[jUѶK c x1ƣQ[U&.&lE"} Q\E1+uw>.,czbQ!g:0rje’"D#cuCXȉ4ՖK(KP|d\1&8{9rL^ K9+ceޞ\ d5ȧtOߊVB>NX)HC(<|1L^9Kb^;-sö!`0ݘ/l+1L#B8U ֕&*?6{դ}Y(INBKhx*MOemT.a~.E zG)j<Qhw{5¾9{`um  C5Q}llK`[b6 |\!O*/m} 9Vt*E":v1e{KK{s`>3X: P/%d2ؑHͶ4{W\hxivoqcU!m|xF^jc5?7Ua(Pnʬ^Cv'A$ƽŇKA`d[_o{/EZ~'*"ȜH*Duƃ˳bJg^raͭ̍*Pu*9bJ_ ;3IJ Pc Nh1=s\ T5҃F<pbq\X%dÜu-ssh x9| kܸ98V#%æ~c ՟Sxa#WS%EasU _" g`֗<eq1.cMʂn6%%<@fpKQ31pxFP>TU?!$VA`Rc1wM "U8V!5> =҆#xɮ]U`w۸ہ :Xxr\q>`:F_Ѫ2)sCjWtHɩhiS+1d!~\<}LY aBRJ@ѥuȑz.# 3tlۋ mوb X^/) &5𥩜]cxo\{n9 uf^xb$2[%6+&掔n2lG>+;gw[wL X1æ>;W$*Ha+9| !80\+ xJ0aJ;ݶ]˒H#ch;)3a0;j`}4JZg}:0[j2#zeVfR9~ :]'##^^Zs`uhqh F% [u۸"׬'XS~+(f?TT)*qyqBt(k:_TX%V)'3J17>\`U`\R֩bJgTlѪyMEYdECaMV&"FhQ|![gIK v~(Jc%* [dI368fp*}Dscǻ3V]dt*?Ű~{N_w7p686~ =WBX"X rr솜[݊/V+jPIj`+C6["Lع`C"cK>HC.}Ε00e>& 2䯫vVj1c$ i'2Sn-55YCrE~b>|Ď7O~mbIapuhm9| 3Q :WUq-S)UNvk6n읨gQ`LRlcxkB$_DH |08W!2e_ʿd}xlzUܟlNDUj>zƖݗ&!jC`@ qэ+sheJm RD3Q{){b Jatdq>*):Wh#O?7] :OVGL.xllTWoqqeC2p;Nd& RhT#K QCppS@pȎ@#gƖ8s\B~Is 2sӧ$6><+c~݊lz1H[E'6/clQ.I`AƣWOlҷ&5fH<=$ÙjuƖ}(Pn*WF|\A\-2) ' RG킖|h?āUŔy[j_ӂ~ A弆^bDyz8ǖQ68`jXcsK58,?YP5䜭ne9YF:oTEf3Ja\cBzn~+CX|,lWB'9DcMu|~z+wG q|-bb^HcUX }rc褏1P[;psE.-m[‹y??ڵElq*E< ^X9ۧ@Z +z7$ "i8U 7bSeo'ki?I+/ޕ,EJ6q23Ƌ hRyl[ERnYʼn+yxW:DLGI ҘJ'[JP&o$8Tš*I>OwCUE0Pu BJwDYM9ơ*J%.E/z<"UaSVs10+nOOyPH?>~to?X5Hk)*9NCNI={:%e+Nݰ 4 =0"__s}ӈܗ؋H,G/+b5 ;csuwF ?eqe9cxY\EzlAgm,q5ω,`w-׋CL;B_[k&yYO| bWY*L9r~tW2,Ud>/_0nZ1"o x` cGCtg^wZd% L´ ;Q헤o/es Giʊt{uൺ" s7_EOcS7Kݼ4KݽԃK[`]hdqT-(l8E` &2޴<>ۦkZYO!,MD/a==7tӵts<)oDS[p%6 -bCa /E&T!2kGZ8gMbɪoz}IaVBn^XLe6LJz]zŸʣameYr̻FO<ž[0C2B; bmݴ,qa~sl /34Cy3-Qv]8d(.G-;E 5UK&ZYu]yzibK~`=FD @K¦#M[2xKP("qQS?3Y VUI@|b2K6gb'PXQ{}e_UG"`.꺨^O.ϣjRꪚk[u^zHu2rZdvj?Y:H+m5)uluaf#V ǣ?eKs(!M9x;s55@Z\4]4%[s_8׏(l Y%ƣ秬 O` Ŋe5eFj3_lIg]DԻ$(3R8͟D,-E] xk,zD0l-e6]ۂ7 uv17Aʹ %lE}W\ix+Yהs߹(?"s] ]t nw2v>f:ch`InN.N( ]TySF'ȎǧHUzjsF};!a e]Vo|Be`Q?}FɊ&XRNf"i`w7r[bˈVWmE<{wG9!>}{v  ʬݗ-is-6f%3eSU+~a.ƓY]]g- [L X gRfѠ "q{I@|})1oif4jyS\df9 1ĸgC@6E]{EFI ~d@2'!KvE:4⡽r)j>?lZ~2+_]o~}=49Gz^jmׂkMVXAҨN__htOA7˳E7WЋQTEҗOs/hLe}|dDY*)=:v-Y_,"h<|dzEC}=$/{9Gokh57o VUⱨp pB4۳G̣ XN&%xN[8j(;#D2~FCc*vܶ4ږrDS̪ku21Q,w+%oCh_j(J$Bi:7\[0onCDdj*-^[qv5KW#1~NDm:4J oLGe.ۏ[D@p\n94TьeR{#rITfOgh[Qp_!2`ȕ7hm[M=f̊j̦CꖭRѢҾbEr'čX漍ҊHSN]Dv]T=VQ-~+ص5ŭ EEVk64.!TsEL@i3}!>ߎ̦Y! ]E0Hhm-+,s6e5@bVzDR#r}5I}Lo?~j`ʥ]k@4[CG ojOKW֒Ȫ-v`&S*ז%3M)dY%zə)򷄁};<Yfe#nda0WkۂD|^j_aC+(8*Okk`QyY-eS2y$iDwߵjQ<+ fsX"Y/4YI@iGu| :#րE0!앫jqTu_Dz /Kǰ w/hg( #fM 6t-nyv-)γ^65WP( iw{xXo}b(6sl0ݱA\mUm{Wx.^y"wzwlmǯZXx5}?ɫKɏYe[SmZѮ$=\g{MY-]KjvlyUV4_0TA *<FS9O|==kn-MoOX+f זA`4Y lB;4,ۉ=00kXQhY1n\6&( AZ%op*یx7L>[P Vk$\P"MCIKJj756-,YC$HSEQ|ْ`ooIwm\{_t0Zp@Wd=ڒ^%TĖӢf}ܒXSɩn*mwg w@awtUOp^W=7YK, ڳ2 ,QFzዼ ]JDQKb2 SE M LYV>x˜mN> whY vDf k_]"I VU9BEK#m? Y ca-wcy97&v Y,`6欛C"nM=OT^-h#GB* Sۀ \cSWTUcMk_а ST:[TU%&ъV5hg-]UUF\ipE@U+@ڕX UKWޫ"Obe@ى0ʀ|DU㑣UZFTpxvE*Ǔ {2{2kd{Us{3/;Oñ1-k h̽@=h]+GnO{^N gjm4~{yĖe[{6 y{7{ղ&P{7=t56bLяn g&G*ON[o7Z"c6tE{|Ho2@B!Y6!pg1YaUSN,Xў,P\~e{6 y$>R1z&Hr}6e 6a/F߻q]lBt& kz84LJxW췟.ZCX!ĸ-pQ&LoDuv17_& RzX:Qn[nh#8 ; ,E{<4{g`)`eILd8ޫa`bHuM'^Az={.5-Kw$I!{.v3vGLWѢIp_ET3BxwF6I} [d͛E Krm$;i.RiڡҨmږD!9r>Mt|I/iuM8Վ 0:{/uzXO$zSw8@p)Lkճ}_T0@?I@O^|d3<EeU(e!bZUoʉ -YPH`w9kpY,ۥA}xwo}8<.tpNο?c~'IwdPGepC^CGcݮtxw'ǥC]#v=.pa:nD֨2ꈥס;P.QC#FCz:t)7YBDk[pgr:5NZ:f뽶@xz8B[VBy1k6J"__P8VWƵ Ps^|Xvs_ौC̷Q1ߵ;skhtlZޮ Vh~褼mVБrK,}WchжzȂM 5K钫$]SwˮϩRZ9;::vQ8md~PzQډJΘp8pO| g-=ߑu3uk~nĄsNN8G73!CZ[VC9q+kk.N|c~qK#7`Ȉע{%4^Wf54ۏ7O;rrvʯ d 6ZIQ[o:uq~8Y8Yn{݋FBs:ۍHr:ZrFj[1ܱ:yAՐ/E>\>w4$qK\'їVc("F;{i5P p$ l("E.$Ug80^ԎG8.3Hd2GyMS?D3/ B?!hv VkMw#\9IeT@䞒 1^ +G 9s>ޓidXcbo?*l9KR{tl4ڞQn?atO6t@cj,I(DVW!-{Y.XaFL­9R1T5(q`^Lx/Ӭhs]rۄ:CRC./%i`8&c&Hwj܇g۾Ei.ل[™ UQ*\ʃJ12FXo#wr f"O PEo/20짳u1S^\4G .[~ 0Yl'2:@U$MQ6-}cA(wRk. "ev ےuv "S-CGI VqxJ-`uU͠|4nfi̠;l:tme_ȽϾ-]W՗$,Y<%LFyϷzZCkԬsLk`~Lq҄8&=<;&OARs3nl0CM\9Q<k-\.K1PiD$88Ռ)Uq QޗtZT!'GNs(2cF&Xkz@xTJFJTA‚  ?@p'G||NӛVQu /xp:+֎ OEk'8/f2,ױ]77'|yx2yJ(7[~{}ߡKAr 2J?St &{LH.rLj ,̧,u=Ƕ=dgVr#/Ncr.O 54M(l>Q |VFqҬzKb+\F.n`3O 7`[-lpkzB }㋹{M V$s]NbVFw]A JMIgByVʖ e[ʶ'=P#lBWmӧlNeUkuc]}snӻ|Z|{ZH;Rʷ Z&ڂPk{BjH2j?P{GB-u u ٞPy:;lAL#ݑPw BeB-'{ގz[/oA=w$߂`` B Gh#Btm Ie o/H5w"B>ᱪP{3^=˻Wl0ũtl&n2/t:rn9pO!t衷萤$i!] Ќ84[ԲמC}_R6;=fis北}+\(Ň~ hF "+Sk6[f~/{hP˙cRràh0ȟ͕Hrх3+:jnv]4ϲ 0B1HZq*},q&OyY ~Y.:]8YuԦPnѯ.Z>_ ,!r40"/0t9kfh5E|Ni<# !t64}3t*՗'ri?Q J 5mNQ6$ޢG*@pnӲ"[nF@T[ѣ o3 -g<9a4AWMag0X&-OݭW::o&üS%3y"ܬNjsX Ρ-NK78-O0*Г ALs@7ПY#lC;,{8|[ر15-ˋhWFsAÿTџuwEp=ˁgd4EX`{TbWH<0Y=kB=cV>0`"UBO:d M"YY'g@LR gUcS О8Qe[^BzjbV<ʓ,)H6n#.&p\D?UV/)*}VgͪU/%fW?,Ψ(O-0V.(*DXRTdpjqb/-8t]2f7W%JSl|z)u;4:C__wE Tp8l"f ) E3--O2!蠘MLg,ie:omqf"(k 2(5kC i5Iy?)g!?h%v " |"X0bšD6D>=@-z} Vh )^0|Az*i`L X ao5k@iveT~jObTDfk,${؇Nf b=7Q^=؃_g$Jy?C5e(rKgiA2-'$ʽ=y#lGy*aI`#XGP*I?USTw{FyHW[b$g}WF9[[x[4]q]E8UF9T/ǧ tzꇏ<A{P::ApQGS ,q]*NAׯJ2ӢHTm:F[Y6"g9wyŔ%j&z+mHE:)o?Ft3ShZڐ&Rbi\hbee\fzphI]T$ئ*-v0Z0=zb}]hJ=Hʆ^jF- ZDS[J6,ʧu+?nJCV[ $Gxߠ=>C)]I @FL9"w7TqN ;e2ө᪥4֯:Ŭ^DTbuQCĭ;$e4}4pTi󢔉?TS݂ч%5 $)jn\VK ޞ&uKcmhc"+֖I=2烞d')FwNqqJ qOU?7CzߘO/<1"mY,,Z.˥a0rArdE\IF>](:Zjâd^%PYy_Trjg о`Pf b)uR*6O"AFbU'Gth :;1):ιj:GZp4'+j-*Ǘ?_\ƹ{>9k 믣2@c׎,qo~L0ޱS1˵צe]0%`~MdY)r=ބ>گXaSEX׺`տ! eEUFr̂ yZN<&)_Śt9E! DYcb+lCTr--YhJGqp 磑"+Ppxvl~h6uX/G+Ϣ6WU]/Q\;,|w4%"p;P&p\E:C"f =oѾ{-5I8! H.õ+2G$M|S`ӿW!i x iHԱo R{>|b0?pqh 0RZ?%NU<&׽-&Md8K)(qK6IB<6krok0~*A{g D sȻ渎_tf/c#jh,B7j47hJ5 {_gU_~os|A+y^Z𷃔_|SFa{oos>ILπ*ehPLpheGp[?R$`L,.(' ؛h!90hȥYHIu1(Ge>R$ $tς2E=Ǥ7[x07*YI$NsBJ@)h@CiU惦Ya+y,8VXh be!nrUKg?HS܂ #~ա~tmrzE?{ۇ‚c ?ExU];%@RJnP"Ϟ&Ӽsaϵ/ Lb3O]t"DO[3]:N3ܛL)(cS4)Np8ވ*gKXs][%ǴBPԢ iDz-kYnXP|!z^C2rb>jU,E_MekQ_LqGI`&lrr)b^2 ĹNJ "@Ena\h+@Tt6լ+{if(0 $5ӊA_Ԉ/d%E)^Y~Me:3F[kUTIңAĎ" f~ i{},0ڹT.It#B/ R(P$D V2 Gǰ,8*X)t9sBʕ^Ԛ(4_ &gVFCgIД,ͻ/aA2={O7G68 6,p81p N`R-EVZJ,$Ȃc{+U@16K R6tDFyIpx=bi ]ދ~M.VX1Px z%xX!"x90gTn߄Gg Rm EaVO''W>Ci)ln]RLlS3KNp|2_a1U]/gވwG2T" 3sx 5 xd^識@$S$AKlx&aϭT{j5a zQ5([vD1&ap)E1\=`xI &zo?\R;%&v87 tn]/(JszƳ`UBwƘ \m(`,Y|m4hg Ed2Eic5Kd YT0WK%Ȣ(>(ƃ ڋ#s3֋Jmg= z]S&/G1Ib-L ̵ 1B&R r,z1U^^ AiVU`qr#ȣmqp9zD "ZMUt!H'@~ch\M_ǦĹAZюN;}QJqYջCmF$^LK?#gQܘeywYf 6Ih`V^d6Gʪ\ʾIp{zP@&/ =c J0S3q6"iʌS<*t$7? oΔRHJ"v~60rrqT/RBf8Һ\PY)NS3&Xdf{}T?>|mV6.ɔK&DLMc`-4*xZʮGc" ~샳c28+*j_HE4ÕVίm4ʚɳ9Z֐=F_ПK6{hm{I|H#H:Uf&ex!2Pp߆Zmp&K'PaY T5E҉\ɣg;wBz͠nluXY/huQAcBjII јI\!trry?e7޻C6# *d6%Vhm,FFEW{xi{3OMIuS,'Sз9_.V5OR\}8s*cf*OUᷘ^?wǂ]Lz80Vc*J#hQ ?$ư$t@d!&bxA)'~2_P8. aUjjok6>b*ЏK"CNY5"!zI {,8Jy amW,R2+\Y)iwlמ] } W֑)&']fI&B m=|oOɀi ײh]cxGZ@#,B=]IS.CR8d^^L)&WC-#⹱ysI)GT<Q%V@ 0U1`0)挫T"j62ítj8i^Uu5%h4Klɴ+3y1I_~k'{M. ''lb erJ قL.Ov3"Ua`wXg5y𹛀^40}QRԱ Z841,@sٔ,ͻ'3Q$1E [\r.>z|gq /`#ڍY$$l7 KבԳb简d^(>k #K*Iz)1kEtrx[6#%@`8K`7{)ț0&# 8y.bnu,8zE75e`=k 'weW*ijQL R2n">t;/ ]&B 5jM )u鋤 NpG7yJn1 !,ʰ[0-sSaUeER':gWUX)Ձm=>QSX ]m97٧9>H6w=#vL^Q :v$ jH E;%"x:hZoJz~U/M=';⭻d3FL!|s2 R93z%h5x绕]:A٦7+k_Tl!w' mAÖb JbN"iF?LnL0VLfOKfc7 j2vUlf3ʕǎ?AI6|͒crZUW3i ( }ѫfoC _r\R7{n euRv5v䱽4beܚt޵4rd~XBQ4e2;blm3] .mڏn G:;kjeK_4m%C!9ZHl8ȰJ:,A=u{s0Rz_*;s8*6os1x4َ^AW D!QMSp*(O[ࣀU5j~{fǖ:\}6Tx9z>N#8|.M?rh8Z^>J6  +t# UUeSD;"?z *d˟"KtP5g6xc.Y(KީSIp #~#u_Q[੢9#S:sN΁k{E ˥mѪdgPITud%|G[LbČG{_;޺GT]*i|NU{Qd]#:*dZ*$Q[@I wv$'hˁc6U@#] &s\vh]2ZٝQZq4_.D [~ۼ7:FcW7U0K|hitYᱰ}cLt;5_]CS'J`2V sfR{ۑDJzPa=)l\{M1jeSgkgN28u+ ࿬M 8dwډ7YI+|qwL`9 +ý:,šn&NN4"@ci@>~Y!((^ sA_S,)ql@Rn&H3yQXN9`7+Y|J:,@%Z{wm6~'ev4 $t3 hCak_,6/0!q}놑b1/6 ]mEX2D;0)9 \Dvf]a5Vn S/|+66(F3gn#igcVJF͠#aOAcA1e% t[Ц b<˒Ul5̞mȿ)Z%'4g\!σI@!c%89X(8,!VՃp(Gm?Ëq1 CK'T@,܅k7(5).>Z+Ȅjݣؗ7z ~/ PRM%N+*{i."TeSʺ3:Tw4K?fl?x2F{֬ v[雲\kF9˱P@,[3P;ypWBӶouWoֺ!:@Q*u j{N]-.Sbe$ebd,=Izw7!:|ۘ `ڹaL=Bׁ>SizZbrȾGgљ5ŕ^άጄL љS4{O'L#YHy<گ#4}Mje&+8n|hv7>> MKekXД'ŗxEհ[2Z%lP-F ܜ;JpsK ׊b!(pKQQ P@Wz۔5\w'z4(-ڵ~d4q-!>Ԋ=:;8' ¯ZZ): IR(|%&*?d@ZqX!$kygw϶1b6/YcKb1 w)X"gNXuS`hALq{ zP9G* 7* 9i&I9WpF<)+j RCz-B`P EBX@A9A&JBCHj 8PDs+9͠4i8*O,S)i P0b DcxC Y!BrPrࣝ?:xF@b$YZ.# uKpw wh?/[DKcLYؐ(+wC=#C(P5ID:͖FVAp`T)Ap ̜9;8ǸI&%42*Sq&Lpp%FpKbsߴdǥaX OcCRGHeIM<085r ql`{-\q{U xg«Ef;9RĐdJFuŇBj(6c+dBvHRͅ8 |8A )Nq[4"姸O* JOq.帵L2xd ][˪ 22>TFI=uH"OO-1Ix Gtb)%%1Js~u֒L-`)Zv'U-GVq D5ޥ&"NIn}b)!zɵJ2ɥӑ)1⎙{v@֫dٸŚ6anj:f)[ =;Q΁L0:MNԃpν#)X1!%1](-i8zBi\wGfSҠڏP]hG(?)ˍџ~PT# "p#\R>cMTZ<%)о$DIw-[Tt׃+[>nVvJC h3jK[W eE-] s|K*d ̀`HPVz)O>4Ru]}1ZLk L*LU`AfලZΎ``i;$ iUѫ0aa"`i:#W-=ͯEQIva~}g80wMuUbӑG|Jŀd 5mȮpm[UŊBՁ)*K8J=화zpm\ZzIkf\vId<=ViLmjOj"joR /Zƚ-3G<0U3fxwP=4Q줦MZީ^FemM.ԃiY`Rs`C>PTn9c`lɪրL#^_cJnP`29[3YF?L`J-8pWƍbO3 Ŷ= B[>L:}tӌ*Kb /E<D`wŝrO~dK7{5[DX"Y[cPzy;<(:099 |=B0$ݫ_1~{*Ɏ~qj0wXwg!\v(IL=L8I8:L).S_{N5jĨM`Z"[?Qzp)]Q? Ȫ1Od^a|:TFxa= 37R0JFK6`ZM[5 Xˣ&\ղXb[K+7Xu@Ԁx= kV7˜}oڼT-RuƤL]aw`|pM-糐"+tuPu},|]n ڏo!GZcBO+(GۥrMU:a}/dEߡ.F/" W0)=rƒ] =8cyݡïyJ{)ހ k|M9T,8(_.@$oÞA'5T,7FY8IhܼhGo%ʛiYAĘ /-8Xo4UTD0z/N9^`$5t4h]gѬX@U|ZэLdBkUFK /BK%C{QTF8(h4uѢ(ߏ6;ΠrQu)ςfGyK/g5܂: ó_Yޗ8=o_W.|Wg?|z?ޕRCe(zpVL!PgjJ'+ӊX%6~s7X=R97'=g)oZ>-,B<1 ?=Št<atӨ(o9m'O 3&GCHR]Nqt6猆 "Eu|Lm&y3Ii)}\S R iDh&~%X|^|(Y?/N n:д!Fcy~=l^DPOt,doC-/Znhd$D iOgik1}O0V@waQs@X(Y 0u?⭯b φ֫&}.J;(FzdnD?qV!@׾͝c%&~mR!YZoHzx^ Mk@s|N:`.fɞ_DC9s-l8,+=Aro^nG,,diiwb;t-p_*/,Z ,B~<b;~/@+?q1 >?V3K6ђQ̳PpVK0'eFkJ-]L*STTו͡UDXNzdzdm6%!< *.ICFui3!t4t@d'} }DW^hHCF2` %b"M4\7!o_U!M"=J_{`c׋F5Ÿ&N'ٓ.853AO9/7w>uRCb.l *Fx-5ûۉN5 ZkÁߏ5.5Bn>Cx ך7e[Ahh0^Qoey irkvW ֬xpg.Fh|sg6|*<>sY|[d~_u~8b}ȟ\L{\V +w_x5h' u >-ixkF.q=bbK^l>ڭ ثwV,>+9s_ PȒ'Wx}\}[ZlI淃l6Kg%dP/\?혟tݟW.ji7Ҙ(1XO)b ߅].go0ŵAxCgm8=r{aWe~'_e obUHMfuӼ e.5!<f,2Zk_ǣaf]4^0CZж1ͺ@嗲RG&p7PJݵc7*&I ȧ>Ymӏh+_|t1|rhTmeh,eY x@痘}(/(ɰ (WiAK2 LX kNuB#DZgL(a #%0kNى|TgSu&'s";MAv䱱|/Fȷ> y^}-@yJnNɗɈ݃m%re vf-b:QU;QU;QU;QU;QjiN눧u_;#O4k*x!4pzDs5AZi׉HPIbwD(c4vm'BAFn΅|*L(aę %Loą . S<]5ߡZ\I"OIBtQaDkΤ3%s)Q)|s:A]N|!Y X=fnqn|uh~m{[Θzm-cʧ`OC[ q5ɗEIUl~ѯyb[qem !H<>I1(}q{ipnXML /H WWKͽDK'bB< G0r I ?#HؓmxB]/)j/KI:ToTPT !cFbQkSEgZilJ%7rgu" w޳74A8'ؽMvon7')n60<&$]cjV%݉}S V1l? /N͆Cm eZ"8S= $Fpvozl+1"|ٱU0rId6e.ޝ?Y[Y/~er.y>&c<&6"x`ųCҵཨ@EGz3g|aC$ޟj.0ٳRN"xIcD} _эe}퓘l) 忛 Cm9 t) m"{WUL?J~X_Y.'QuB k,4iφdݫ!&ޗtmrC%oEF6S HW߿y,7|"j7g 5n%J tPթ[g ;kcb:L9{; ,R}Bosm;:o6Z2'cҊrXdLd /ZlK¿f'&8/W^B$+}qȲRM. ߞs||ɿvswK]+?} u>}%Jrmh(⤣10Q{k?EGw/׿Mӗ [=M5f@5|xC{IV %!Bq ϖ 8BуG>rs!.Ԧ)nxʸ-“@vVfef"2#=qUCNo&>WM$q g2%R'Gъ!@t!H J[`8|ⷐpud0F@w_BTTHIBu CJњ װ< I_ݠ'T ( =k8d_d[E!YbDQऋX)KiKDz_îpٗ ^J/@_ie}?F0\q^H|Qv]O?+oĿiOӿ~3E_>Ϧ{韽yrʞ@H&2 % -<8G4Mr[0NE<[.mm%*#H A9+I۔00 |~1泥ylZ#]8`gC(4eȀ#=q!.GtW[ʒ;m J5p14; 9~tfTme?=-bp"Vp)U .ZRRpVm> !x}2T ; qfk-߹9ZR3>C2RO^HhdiMd.q%EG1zXȧb]Qrx>↭$xv5\e(&Ӂx ȧl:xM7jE("/8J֛8#2cx ǑO辎8fuG/ h-e~7WRD!p^G"$nqF}|}CE" p@]-&n/wO]Z3b`XN88g8^aEǰ# ϤN<"jbz`0o6,@d;}" yD6lk/9k!җLiMD0b:z'[槸~\E>dl xQd@F˪+R cZRv/8(Q"$܅/v9x"2N6GAOֻHۘȾX}_dac+.IP2}q6+/ר?C>[N]M+U&KǯU^U*])$)sxZՎ"E|N /mTߗsJ.B'ȧ_1zk]Dw-ɣ5Y-׶GhTÉkù L}6?}Cb-[?f|mD4PCXw㖵n*Aysps$F:4mp0ƽ)h=qSXyVͷ] ^U䭨 Ng5owLHŖBQ' ^X:&J\ تOh#XFODH1Q&}S#WYxG+ -sֳ.zdu" y +6ʼ[Md7 ͫan 1) hrK_(WZC>u[lW/Գ)Px_ᨒ v0kT'L%aK LjӉ*K|EWDcU5~Uiz*觴2&7/Mn0Ct='`>vmz}U3z#IQmǢf#0U2&Obz|n;"d11?³VPf+d X k8Me:%Ѧ(&Mt8V \.ʔn`Aq*P)fCI q'LKK:nwZd/9 ,S!:ynѤKEO։4U׬ם:Xm8R, -W+ 4Ә& g fK? (,G h8 >kUk$-8%+%z:,gEL8eJgO"ETDODU4۰e*X[<SYAVLHa|l_T|g-1+z@FKb~1?l+ ?*ffo]'aWs:ej7ޗpR߈U빤D%a D'mH0>6!6ȧ+[r\Řר&[o7g @ɮd{C #&Ccd=d[7:*J=2!qƎ͈ pioYFu}#9!+Ao#83g1PD2 T,/FO\ze]hVVe,8s^}L=s5f/43l_{}C=<[ruML@*1{˃$:AdJ a!(%WQ3-y߼E>Tt(+b7q3Ӯg,QT1[P1ȧZq&#x\[rL!5B,K+bH‹1W^#|*V^dw3U༦9 Zݼ]iE&s@FÙB[Jn6J-f gDtFT/78uISWU8Z(㈱$l5bupɜ;8u}E>rKgM|{xg`ES (_5F5۵eO_в]mm=r%Z'2K-XҼ6X'fuUrc|oTNJӥ==|*ґi%IY͔GrE㪰i5X=p(t¯y *@Ob5q1T=%BИn0ߖ.dNTgȀ%GE>rKrbïNwP5 2WYzAƩ_" xu9b7&8 Рd4U&YkNJYxҚ\Ja7FOėhLt%17?#F@SkBWRJ 'RIу]G.E>rU4)2nn-o@f3j@F㪤`*0ePX`5Y̲֏c"5W%ԻBrDoBA x9|*zU9ׅF\Q"<ɧfPd4!J]of+r h8%t-lQ[uH[RU8yV:EaD`C:$i{H_-׻-*/ŮLlɸZXua/b<3-ʹ|Nr h89Ds erLza\.j;W·mSH`R G/ 6ÖFT[6V'p?p'wiTSQ:}ݤڨ75vKX[ Yfa=Ll7DE8V€Z?j!cȧ+ꩉۇaCϷn*#ݓ"ZQlȗVU\sƽY3zJF ~v- ތ֊ Fä$D ' `A)/" hg0>$"3Evctmcך e;g y},W%qit-j_FP˵qL>%ѰYL3 !5m w?`_pWx(o} b3f6n}@FS%cK,XŶ>~]nL%U8, HPQOl)Rq̧x˧jؙT|Ɣ-q[Tqv΋2NRTDEErtA2nT啄 ̣W,sEyoȧjh<.x4z7 rξȮn'N#rhU,4"ʮh%b~1 |"rOfi1J?dNl'pJVD?"C&^_$T]G7vG5]1 ϣWeЭ~Wu Ǧ+_˿_먺ӆk2-@R6KW4$&ASDr3_X49 2ڠl$k0Mkk,O?JG4Dp;8e+y.-5j8bÙͦb[)-a3yjj? UӮS+dh$Ĭך 'L)mӉyۊCNCz;5TZ ~58} Ar6G "ؾ¾L/pC_7DteEmaxkpK޵47#Xc9g'>w8P8H6IIwdQd$ETE{,d |d"yskX<< }VMl>ؼ^9dZA{ܨ_RL`ED.``[=5:3$V"؆X9ɜ4vE|"v͠bX+yF12b3i,b0(+V1:\wLHAb-+>_z87Wu~N=cȊ1;2`&^˨Gl 'pa/2gړ~9׻~9slby~;ԑC|7lyn*ʯܒ:DײUœ8QY5!eqJkծE`.vTaJbb\Sڸx+XmF]КK({.v7\E~}u ?Le2!2Edcza2֚13x-Y2Dj&SC/½n84S7x/=اUК* "ij0ފZKnحIr$A_;ʇbi{yJ\ew!klT"uru+Cqlc*wک cL%DI3&F纁H );q7fa⩩&.͉Cç8OqNq"LOr ׉3ӊٯR(:x|׀]] a`{Twƺ|H/Cg͞v5{";=(ʺ?O&gvͼږdYalwYSk]Y\M:xL CX/pFJBS5Eq<W7wt輈w ~9OOζ K *7 A{3悮1x^N]~{OgϜ9 2}|$2K+yxrzr ~RHp+wӘ^7/{a|I>"&@d?f`O\ |}81ZؐmjA5S愅V +-3Ǎ'TAf`ґ74ɱ!e_QJ;\PfSti(cǍ'\tND3EA5‘[armkv%J'n>[P'LeF9!Jw EsbcCI[pJ viћkC<4#ICwzl2Ώmb98]T*vVӠW-\Cۘkη-,ie}3OL!j"۬C|1hz9ءaG>𳔧?_`n .F-(.pO ̚DHS19,a``4'\PX/A.B17Ĉx]Eejꍑb.Sу.c jS>F'EpitR q/A,a G1f_> gB%RaPi;%u`MVAq@CAJKV;J b0x,3+P%l/ 3z04VP^NdT,'^(*6>?o2?5ӂxZ0X=KNʄ(DJKYjI',֋U]S'.F/NQ`$fۆD)"UJ j.oy g?>T_0LJJ^*.*KWhM-76E5urť]>p3a[7mLEt2R;]e!DjNM-* QKk'@,aٵ :{oM Tmʗ#TtZ#ņ+b\񳽧J^$=>NQ+ bL۬*Ww:TwR$??NPxVE$-HV9$٘+0ϝdE"3zc?`6!`nWEYˤjV#\mxkWؘWяWFjcPxfrY2vҞ^(:.Fz,KVE[UDVm59t=ipũVnJw IrT P)K)p] gP/uNaRSo|~CUj#G%32Yc DUP0ayh5юG߉vD1xы"ůRҸڝup`JLL/Ìس!Drv/[LQTsu$t&`fr z\L7R]q0߁TnT$?kfLhX@ ~,d)Wr;֫Ҍ)7_={NHP@kR IJ/lם|-лEqLǟ_)ӷYv&1ҋ*P< '9ñsᦷC438gh-F/ۻv \aq^R(g! ]g[~{m*)U/W}יϊVHPң BBcà,7NԷڱ15`;*u~Lʆ c݀0%GZ0вDeuNY&v$֣ gOĄ%ިGlkzn 狌9  zS )!}Bf".PIgH´+y DŽIH?mUy$sV(JSe02b5 6o_r }]E1ӵ nʌ+-ޣj$eR]"'.e)kiB-<$/$1"PѢȴ΅DcFnjbO[ fR { ;]/:yO}\fZ|ť^jGY#B ZT5*9Q6!Or^۔gU@ Z@B%K,UkNz?P$4S?9\|1f~a#|p:bx@$(RCVK=r;(bY@m_vV-J|~i (<堎2&{\{V}פIIQ@,mk/HS)7j*7mS/ռ6a Cr ѫv@aڶ *]+<}{OY-D0NU4ۣ=U#1%]`&j$'[_K3ỳq$Χ$L~b{{ iDɬ]OϬR?cFPKouZJ݆o[-bL`,6)l>~i//9nS3 -gNn ?fz(\9s_%|_&>U|rxLq!̕i|TE.!lb%`#Y".Iʄat?_{8LGNAp~q SUz{ э.^z|#8aG܂*eF6̺th@dVs*?.h\c&ɂ)G ߠp=Ψ]w2x4eA*۹7p8<0KQ(*u\q\.z|K JZTܕ0c*nf rɿk#FG_(;Ȝdxww#&h-5٭8߿U<#⶟A~KM&Ær5 k@؂V[4[zFo*[h[`2 &췒/L)Ǘ1ΎJ^0,8g_c]2 诇;juN7i<]6U6hŸ+^s$Z$M@p֖{,E ̦d:4d6]5Eډv']]#!.e4xqx-f=xیͳΖt pP=OavtqcG-$J?Br=ˑeVPSF抰J6(\ Q .&O_߰{=} B`[/ck'T"{ +^Dƻ6l$"QI;@*elruOȾsuYY-0y 1<08Z"ư]Dydhq4#k%x8uuRґj.XヷUmPurb0V1l[ܬ>ԸZ]f-+^.jֲZ|[§*n!~%iOm*X $^s"3Jhۺq9} 6fA < \)m>۷rr6pN:/N\?" ͦ38| [UelIF]]i/FLDuac>)yqMvLPghAI8Fp$XFZXLqoX )D@bcm2FsZ}cwWeG[7^T0 F@Q.^T1&by=WosEI\00JFO-mz:Xs`c\6De>R-L\Uɯ bUn3VbߠDJM։qg8w0Y-2'ë'5>B zIlw6Ct[" F(I"Ҵ/A`Zص.] uQ{~G3_ }CV3?w#F=85|wANRs(l>*%q+$S;7 NTuV Ao\q1|Λ܍|K)>/K/5s.Ƶl\_A6I1mEj3^AmYbX hE]@xArHAJ2+?BP#gfNh N+◶;c#( {oz٣էFyiQ cvC("(;qf61"Tk<[8v.o423g?x51huUps@ESz>xFN.XvN%=zd|q Y5Yuh3k9I($Ɖdw<*gT?#Rqea#OT͂ *NVL$| {}x~zƫYy 䁭^(.A͒hVҁ #iٙDqGEJϷbL?z+C1tn]܎4/skc{m0ń@p;Ֆ8g8^; ܑ33_{'cEOy33A뵲wXl sCM[r?3ScΤ9ؠGȃn-(bT Q\Fgfa @T r A]>t6ַ謩nakcҚ33CMy˸SDuJ -pMёy33a祫2\r7@nz7oC'UA60m^.p~5?uylF;aJa>!C s[5}

uXv8yf-PI-o<*{Ff N!-Cq58D2@&'(陽Rgf ^U=TP'hdf N7ҍPID s*,0D ]HZ)=!SJ1:A#3s2J_(h{MVղX zEJj/@mݷ֖BxtUTI,\B9c Tbc!Y>H>ИHMS1$ #,9^qPEρ](tj5[:ƪf9{$'hdf:jV1 ' EjNj||@``MR&ꔜ9B Xt BB̗<8DwރQ7O̿ZF}NUEU޷Uj,@Nȁ*{gN!<-f8SɈʽޟ<3;|' > 0$C7F a9NWb%S{FV0:@˥9L9LRzXC(mGAEŅ zxUtn4*啩̼0+ŒeutrY142gD~;ԍ)cȁ'Y` Q:ɈRk#Ƃ´Jl 5a6 t9Jw/iXSހAy9j֚1K(*w̙pGE18 Txf .ZQy33aF(D?|9WߛoU e#-n<Ɗ\5nII4 kbq 5423gU]ζuvxDzQ0ENN)9{#Lq` ᑥ-4mV殽6 h.yf3I:\$AL"eHI-vg3^%- VbU0JS?&N7Ɓm/a Xrc zAZoWjQ,\l\{|o18,iJ) 5%I^ٹWaIŸcenE Qi7"Ǿwa8XQ P&ye&ʄ4;tLRHRhGb>L\z28Al4c .Rs Ej P|rM{@¬`棔6A%am>_nUaG?KjUսnܵrrRq {lg=?:xj?&&R-fwc8S=eI1?v7([zF0Ɓm}M GVx"..M)}7q`'C Sd^pB91&ĐLsXwQw~u`c&PGG'5(ٸܞ(:0}YqEs8,(a.Lj]$Ï1KעU|(gy8ޱy4eI*nM[/1B80q8~ΪOJq QFq%$8e23sYnͪW'908 |/Pa@Ot`ZP[J=0 <GR}GB_`iTD@ X wj@N`m]R)Tɛ-<^b7q`a9b#Ju5NN_wCb*My\м qkOIl>6AXx K-rV7q`p,|X=EȋbmSۣ݆1HQ+cW҃" ǫIT*O'%7Us^g+DjFϚHGa1)˒"QYeS쁐Jc#/HXٽk]Y91}G fY%MvзME1D}Vq|Vo ?ia/;=*J"vEE46) yz[ƍ5W7m bW{3KA^fz:5? ., fXtiT_͛R2VQcz%ھf9Vq_p6`9sǮNJ.FB0*3Fl)f e]./em)ESmUmLIK*XʧL]8)^O#Sh?kFэY_bۣ2bt;M_cm}4WQmy{,خF¯k6%荽{/!q?񼜜lA Ûw0 o[` M?^wݙZh׿c R! @E2ս;G礶_A51"$b'ҺFۗ7wkL"r/O%Z(y3엨1 .n1WDD+=Wɣm|E@ٝM>wA&i_)&aAWg4A"2 {.>bVjv)}tLR?V׫Q@M 41K&,6y(MKw)g }fSmci[d14d\-Fǚ.KeM$"/ \`V,C*NypfH o~OY:WeQ\VgOfvC߂-\4,Q)iy.0xqve1L4\b>=by/+yToqURPbɹDJFG[Wyq@9CbFQ8Qػ1KT9Lغ,nfq[<1w V:}/ Lš]LܺAl {zoS۲idΪT3f[^ω@n:#qq`FxLgfg.2A7YƵr ǣPr|XؓjCj*v 8#Gړ-j.…¹D:^M\Hi&~]6cCf=rNs!1M}4g `e AoZfEKێ8솟XX /zWa$$)u)*>@¨'8{wg}kCȋ/$_ȒŗG_R̖YЪۗ]([ lv]rc{ Z%Lruvj i?v9?{9; 0Bntxq0e@|`R *!U :7 dk)Lϩw4|FwJ+x9oxDQÛ6ֿGy'ݯzVD +…X"ζ99|}OlStEL >^]xzxp 9ΞD 89Z#p݇m"H1wW?`[Qg>R)PEQOh3eedGet>.+}HD&09I`Fy/]eg1"R+ uo?TC/**&gfN`0 n*[Ax4?S%-R`%ZFݨ`9k)s^ؿ^z\>y8դqD]*Wi2xY#7seg* ޑ&;_yVgY']Gw@Tn?\6 [=W^gN?]ha/Fl9Xn{1j>m\89ѳq.JlBWn}OPZ = X !򲧮C!?\vga! g{Bd{@$8}yXh} =c`$4 yNC<1QXc '(y!X2}y!h@czI,f`c10g&3lyAW@{@B B cbS3 n$2Ud@} @sZS$SUTDڤ8:)(I4>B /m}_ !>U#'!iX7 H*qJ2 8 !UC|< ' Ayb0H]76+:Mׂɀ@y.){@<5]oE'Jzinow!*k~>Ϯгs= ڟ-M\au`s~ro5wS$wר&ourV;/J[!7np5“ q7{Zl;[!}b K-Lg5/?ܚJ=;-'8^觖LO'eCk(^?ɇܕ,ρR%oME\u%jjM8C,ͥB 45`70*H4XVW6!:7.Ytcyzf>wynQVe MWxjW2~t}IM4X[^6nZ[ jvэ{^zf^Ueu9X}&U+7/+FgZF9wQn%u^@i/i~˦å} `f7֨9/t\-R l^uQXEYlczO'Sjt泺1oߖ-iN^PQ[OnΟ6wX{]qWug&-<֤WMlmrZ@[ {e-Q(&Nm`ݖy#W3}h7fVw}7(QQt2\%sޙ& :L&r>Z~-HƄɈcKE ~{`nh?9 !~ AQpM'jU)W<4!? !nQ{^m_[(,/bnDc Aާ@@<Ih g :5e,W:qJ0BE6pu!Зf< `LAy;k8)I!{@ ~]S)DB[wu%iLBf(%t@@ !<`V֍a4HȃbD*`Z?%o8 !l@ #*faKE"fR3LAmƺ6IBh:"BĆO#!BwB~B1c'6n k(#cS-Hq@<(~6g:%N0,HlB)!̌$e9&b]}ӆ!= A^^4>B W?>Zy!3Qg><*6<]HoRkWkiӺ?~_ݫ^{?{Ǎ~rv]XQfd;b+v!ͣHlIfzdUJA i) (Xfh-&T'@ 8oC|y\Y(O}Z(1xËoXFrI o(DH⽔ޚ}BsLUɎAy%ow!Ւ`6)8$Pbd{3-.~>I5EDc8XwQąi>-BWq. ifUQB+]|)̍F$o9X" מRe{Pb>=R(Q!E$z I=>-.~8+ % Ò+a$.iާ қkyU?L?Ip! 8ZxB-hDOc*=A<=xbbY%R KaM-ɜd9={W vI:e͊-7W7!J[(Źwr ѯ^nu/ӕVX9LvܫGB,:_2{"'ZD:,vOxؐa>vhsKLZzlkA>ڼ yg }Cix .Tnn >ԌQux,xzqg%N?`Pro$[rQǎ DcLB3ccIIē\x=qh})?URajJp(h<7u8;f~ \7@3P ' `>Gr6/=3T+ Y*o0Sdb'X> XfH (y- PRj, [C"s&S"'47HъP`tM{WCܶIk9 )˶l(ogw%u.T^BzSV&o2-8[A w~y;{Ԅ"!qz?uk:rp<&2oi^;dЂ@ޮjNazjY\&G=cE4P|H̹bD=c޹F 2yF4Rtkq5CX,YDX$N}}A>hz*B+!=yqKw,QLD83MELc2X'Ob !XiyCd3P!ZSY 4?N& l ֧ u |^ڳW“E)H rcx9X` R| f: ;$Z Y&]$; %Ӓ  -sHhU})?\DnF m)I4L:>*;=h6pN{PQH$!;c$1 c20t|d$)UA2&pҪH\B#R؀ʕ.퍕)FfY2Ņ2',^(U*JEH>B 9$aAGA%6WWL@EJ`vXEJB"Rk VK/s,CD (QZ=kZ=hjZ=揇5RT"銣*t%(IW"J$]Vy0+|k3kƈx>:_zr@)^).T*D\a+rUVªiWV*WSPTUVs+rUEbA%8e+V*WX \a+r=9)^*\a+rUV*WSzBUW^zUW^z=E ۷ۺ}[omݾ۷u =A8aF|߻úP۶ #o,M %_>E5w +K5oųt ٮb-EGBs**axI0k0W @*#ŕ0J^ +ʪ0WUi\,U ޽ Ztg`ܝ;{;ݎ^e-IFxV[xSM?J7]k|sF)p;L4/'mhä~<g0ڄH@F[ehtrLܯCO;N&ۑ^޽ZQɽv~߽ rnlo2Y~r~>;8@>ZW]p;λf ,ĝX,G_Ƨqtck>NٜFoQ3U_}=:IRO+ 7S^Zڮ_7H{:(&cV>qyn/ܫG _6wu? Gxڼcg fpbzH5y?F5h:dg;2F)L~ocŌc13k/n1K,-DΎh)㲁5~v:;f7O Tyuokė죵>%~vbqHa”z=MFҊӅ\e?;cXC69tk, ZiwN 즙 j 1&Ek"IЭfod ̟%XV`Y>++ㄛMBATt9=லj"~7٨_H)2,A6 x8kKbKm!CgqD& -Zd V@!W;޾!%9zwnyߓ ]&evn@6=pm(%c(3(RbJnp 5FD)gwr@_ Ԟxg0N;=Sw]3/7oi!"Us8\4߾xqt>#x}Q2+2c@=D48sg+ϜhOq~kIu`d>vne{H;hZ:1'ރiӉA``3;n'z;<5ƂL6i5SgNub ?-S %0:ϩ? =c0{ &$[gA^;݄>ASq4 ]^qʝ„JB˾?lR%>aC٨~z_TQV]ZA•RznyVrE4G+it[;WX^`e@Xx YN9)p0㏮n]i:qqq#uqف';== Jg;;td DkՎDx8Dn|r+a1Hn8Miry~qukoe,k]zw)ڴŶfi:+˳1'ZCGDrDUSk778\v4ҿVz)Iwr;Lcʅ%\+V%\k Zµp%\k Za^dzkvއ8o BG<( 8Hs5'jIК}y{Tv!Gx*vT9k}'NVE:ܨ@%@]FHFC4{d H2FRZi*଀)΍\;,@,L))9T.Be2j$FY$:`鱶6aS^_LO?_sȋ3Sۤ^4dw29Ҳl%T" =Xb#VG#$KS%JQ*d ylKF=L[wVk oeq9(Ӌl*3d@:>} NJWK&kI h^Wy2]TuףAWAyʼn`'xpY3\v q!"K ]e0neHJvS "rg'!!|B.a0q{|ȸ a=wAJ6$~]Y@䣕d$sڒ[S-c8 Jgr,:ǥ{%'i }iǘ*C-UE`< R%rC" D!Ra55I )֝aB?j=sz}.sON!Ꞅv!;lv_p:|E{~sMplx˴G{> ZÙ֒(H,6yDXIiHN:*g`O/{l |EgnkhfCw< *R%ϕ{& hES% e}= O"-;Uc]3>ﴥLBF %"Qu^4|fb&`̐6e,V}41$.Q&G")ypӘ  'n ,-hOctՐߐ㗃%7οMZ}[_)4~rra~ߴlJϿm38$fo"6м2S:f>|1];٭~!sbHC[Rvh)I]T{k'yⱶſ"dѢ{zUD9xEq>ILc^*wij ' _pq3's?ůV{2ۆ$ N`2/kN|ټ}y4s*6aNtflohr^ (a~dgx>f`xfjB78:N(_v7|sy 5;~g{۶WO[0f)<<<3q7jE8s.I%ڢ$!86Eܳt(!FLbdodv>kB3*o~"d45uBH\Ԇ[`oDgk9ocN'o.ƷŃy-|`eh:mTv#{g.n'Be= Z{uCʻQybCC~L׹,hg$z^byV" vYJfx +g80i 3@.n"CdKu\(-Uf/BLN==rgělYL,fj:",? .X4 &|Tg>v/ڃ[8?~s߿5&Wo/@}W>/fr!b}Ie^|n-é LbEIʃ+RSED iK g$I9  SLi!NU N9=`aİT8E2頹S, V`K>&sT:gPkxG 6דo?5.S䉁5^p,2!li1yl&gLRM Mo,?H\.f֐Dၱ4F1I`)$ŤdzMď'%NV{f[d'"ՠHdyn&h} <t "ҧI[cR;ĬW$0^ w̥Jx-lzRméPq!=#)li>߈NL*rr Kr͗b70_XCg!~plhGaδG#!֠$UAkQ%Q3mX"Vy.ϲbJRdΉuj(mTCTT`c8-Po31&V9B7a|GLQ\>@&)(1- b~m<$ϭ`j1 ell#+佹?DyH[9<+>W9*Vdn' :h]nsbBRpqqOI-{$ӡS0:4oϫnz,6.W<ӝV0SPJmŒ6>Mڤ6H8wdn٢:~w-_]޸ϑ'+1Pho "03U& Th+B[O’}&HB kn؏}(FmM^TP1ٚ;[Tު-ۧ.afPmJx۷ӗ! haUWmXeGwUYgt&У[unW̵{|GD _MݺKDȚf7,~^ƚ} Sgg(̕ g֏XWm~7xU|'~ef7"{Z<+t|) Xh6qh$^lWy_"!Mux5Ϋ q7!%dD:Em>Pb#pyk$W@yv5:-uO:JNI\ӕOol zbH9 rJX04)[m͍A, 5I˅BX9mup +!/e԰)@$z#KRg2cm* KI P"!&EYY0lFHm|#x>&Hj>Hj+^erfʜg"rw9F7^/+΂Q ) D\O9mph=e` gv~NR U&)u_Paյ.Xf+4;iƹԼ&Ӭ{U4/FFۘ^\*v24sMtkVv#~LD/aN(EfM$A)!JX& 13 l ,ڑJvz5K)0Bs"o=PU"њ5 w|m6ň#H^GVb"J \ǫ:~uʇﻏaқ|vkoDX"q-Cl5bnG8ęX|6gwNwH3B;fxh?oLwGޞL`Zvet?rNK=fRPšt`hbM A71&ĠtnbM xӔ%̺ekgNi¤K"(ZO%L_.}h\JF¡Ʒ8iDTo\ۈ*{^r|+Ub?F5aR@W) .wccPR@ؖDjbljA*iw ;.* ! LJAB8,59)I IO p>L.8X A@ᦀcc3>2 U @4wyJhPQ")XGjQ Аa?V+m9!2..`n-J6y Fpw{I>e=;uƭtH0WoHkR$,x>q.N<Sti,^8-j?eOc7g?]na~ yj=;#|)bi"::Gz5; 腄dNnSwbFеh$ttR|6_  :mq5iatHzp=[MȯA tgHI6|<%qgI[=AI\Ԇ[`oYMΜ߾Z s*nG2Of\܍oZ0t0]K;|mGߍ3r 7]N3zXAoUOݲnHy7Jv3O,Wph}31+W1J~s?YѬv=^79-J{%h}I:/K2Jl\g)eL,g%Y`"K=]9+Ɓ Oϑ'g0vŁ< w#`i+S|:&l͖*n:YJ?}푋q nr\1+>m"O oZhhG% 8 S`\{p `o._}^]\?^_D]_(U?Ҿ'%CLn=߷[ A׿^/+>< |cm{ "$WFaAqV)"u1JpnLN4ߞDu3{c_>1"a)mw8é )'g4֧HZ"]Ki"2DdFa|7ؒ|kkU<Ԛ+Q =Mm) 5^9+#%E.MFZtBZ6GI>+QGV)> $IưI 6 ,omґjIa:RF\8X4=q"D$M)҂ib0F5KŷTR՛mPa9$[HMLc1J=Y q`9-RLpAM چY雘W~i}N#NqOwAup(@ !upKJ0SH=TAg F8F'tڧ{AiN* bm΄%؁ |k uU8. åԸ蓲*""#dy25&$xrc|m|zI8B9cX#TZ5@$HJ.@=Yf@4P=3btpv?v mk\ j܁rrN,uIYl#=V+!3wFFj/m Nf5lWt0t$D)#I=T#JPY 8-UZ^}8,9nԿFf)/Gϛ=U糪ӥ.nʃ}\ι|#j琙BIRH݉+8Juk4q"!!~"Mb9׉b)\% /(<ъۿs]ٕ{MoS(\P X΁fVdW|bƷ`:Uت"KxZ=j8t3~%ޥBU1b GoEd5ƏfڳhxכYe}~\rݴ Uۏ>d?h]lc=¿Qs7ytNkrogܧhh\y׏;W:[jkeZ I7*?~M`k5  xWe=mէL%OQI4R vXnhNG{ 5 ~4{ 䁒% 0]&KL8rO#겡JZ UO)>ѻKǹA7r:A$qX%}yA!srjuޤc9nOW,u'5Aze7T=a# v: W6]8 DOlP&qxնA=,'"^A:&Z|ЁiiV0B1UT [cR;ĬW逰)A29Hys)y->cBm|u!̹䏕wF3 z%RdeثG}vLst!bNn T86Pt Sw =b^ bޣWTuS>[xm*msqLd3s,:O9(6II I_K,qJL5a76 8UV5 g- ٚ~ֆnYV|8rU i6+u|+ЊR ދR h>dOZo"g|\ɧl+WJnFbl2'RSTiJPs`ΉKu(mYyv M'_=u`R(ڄ-m|I#FzzJB~xժhYQZOvҵ Lu(MKz~H2`2Ç ݞ@2$2xD)D* *0wHsIMNףevI $rv"ԯ{]?hwo~>Y36~8Wyׯ^m_bޜgy'AۄO:JabIޕ\iÀz3!bD2!)J}_unR+@QwWU~O)A/ \&#&,sG-%㌬o }SK0Lx <|7Y2p^nwLWEi՟Uhӯ)8'I#X1`~nTo4MȤ00CKn󉀌!la% m!2 M&\ө\D9DŽRn&x9)CGEdQ0A[R.k%H!$KGKvQtvN1<۝nUް'}:q2Y_ў)k͹3sDo>bma)9 Xy7#PQv&fۂE}6FkV@ K!pˣ>m>bH5bgAYMTh(GZ,UyGL%;-}?A!c07‭iU"bK(F[ i!% 3Bs4>ڠH8S=7Sbz;^^ުM]'p8敯#NmT FZvj|O}u ['Pg,bG;uCm]TW4hYjУ0 tkVQOk) ;F B\MU,~-,Hxi׾HbVL7݊[1[1ܸ7,<JUP+AEObke 邏R\t"h-Cbu})5(kvlG;ya;]dig[vZ*=tO-Dφ;wuN7馻:ջ:J0!:5H1|%"QBو0(qlGGqDZ?} IZf:< Z=*FޠP2@q?&L(ł9 ˭.":$MeEE>dψlfJ` ZGK˅PHx3jg>LLj9+j$>I|q/ CLZvt@_ߪWɳ^3X{\/gd[ͽ<, C"41fn tS )@ʪ!$Hz N3=8QR {>HZ#=8 c;!@mȥyk4Mk $ƨ|vδB\uc f#@jL` p0#LEŃ7oa%q_NOw/V搾X]b֝~~oY%A˶v-)F4feU4v\BUQ+Fn[4rqܝgITT|ԩz^T}!$)W]!>a]%^W,mcB,dV\tj[&w/$yN?HqeGp6?Ʒn;d5}SgL.Z ~p,mBDF1v*i>ڳKT!L}$aX62 gO=1Bc9䢍<'&j7uBcs>ү6Ctւ=[Hu[ `4&=Cm24l!hrn%MHK-Z2prǒ38N9ձπ9`ڂkR۞Q-:}ሷ &zxIʸnZ?,iՍs|ZJ1褥7K110ZԪh[r_Z$" X/ͪrV_UJ/ָb4~g@ԄK*㑅H>HԔ1тFQ4`pHnB PAW{_f?r9>ۏ*YGn/S<e%41wh5Za@QAsglrD`p'!&X|sጞ2bٸ|w 5xiFPɳ9y&sqxDFCt!Nz:Nz:NӮT9cu.mK[ֹu.mѳhqԡ=6ohs`_jr,,Eh\¤yQa1XkőaPgrQ8'**1]*ua\,{Fw,Ehgeiu-ˏ_Ƒ4|(dGVbpD zv4@B*Ҟ+v:=7IF/hk&a?\~ь▱ō,nHs)np0%H[]D‰uX2 IbefmtJ7A0ɵ r B($<՞2#,RFXp4zdZk-]x~k&HQgߴR&׷UҧjkbXf,Xz%b= 3U9K,S(eKagb-VڙA G-5TD(٦c:3XRHAN:#A"mc<=0ۻ-ػ?g{wbﮏц]TsVL| x$ )2LwZ D佖豉`Pј:+{v`: {uep}?4Hٿu0(,QҒev%mwr֘tj\0RbC,n6hF6=*3j[WEn~Hu}{J9z5[7k|O3qj9Ere%֜O~O],|SsJH{## ͫ92~p~p#-[;a }`otdi q7 [@jVX owӋt 's lea?=v4koYh6l I>ZVrw1roЛkCR5jxT_> KKB9UaaR"cٷy[sǻ.Gr\ƊP2zE v@(#d+S.}Ԗx93Y{)Ck^At]o&~f+I==ԥhd6D%} yIEt$LEt!"'U{A|ʾ \H☁ )f:$ rZBx`#m.,cY^i8XG!1z ",iP΁:KFbV2V̒%-KZvւP4t. ロ#gkWSp06K(٪Iy -ޭhjn+pJK-E!Q6} E)AQE!D&uUH (WKM)@8FҎY | \K8GK< >ߢ U{uŷx Z=XGhݵ>fRѼN#;n_K!Ŏt0?Y&f.)5J 0Dl@&)">Zgp a!RW9PI^MR |cC_]V8;"E RrHd 3âҌ;$$"HU' {g1cPl+իЧauMXa {<]S Р҉V{Wj`E( ~A *((!#Jc$R+H.p"&,KX.,H R5d\hɡ}0ŀcDBHOI1FCf`ZTxtLߏLᏝΎM m;ٰ͊I>O 3HR1;Azz;)E$?T!UU;W.c)cA.Q\Nmh+&鎆G.MsSz.pⳲVZE/ZwND`&og ݄ig5Xę'9}ȸ@inTϡįc ^sdjkXuu?|:3jVsgo;E&O$ ejnSwe&pkW)3WٻJ`otdkB9!JΖ@85 %`qA&V!LW_f*~iIb,̲15uEHbjǟR2¸VKdcgmN;ɓ RO勷wOV-a`nͷ {7*R/3~;;`qa!Tד=1ۺnH}7Jv,@h} c(?'dg759-j{%hwI:_6ze[_`/K[Ȣ 0;[g8l‹')P2]6I5$mm-IUZj=],;}{;?&.eܿ_JaMYU`6,?8BTjTIjAzO ?&}ǟާv}C}_~z |fɣx_Ru?}&n[Ubm~=.[5m߫j<5TÅbC,:ɕQ2x`@yPR=O)"~@prO|(Js6<;Zx^qjj!NU .zĈaJqD2jtK)+:t&az7:}ksdh'֜ `[ R[iOQ3'_|jXgiXE7#iݘtJfڢܥ. $JSe<:oUR޼+&ooe_+Ak?#-CO5')5ߕSQ]Q~5wXuԘ hI X3 ,23ɇa)"2IY<ţ%wS]ݬ]ޫwN3w1~~xtx.;vn'u,k;qD1:j<BN}Jljqp0*b1/#s1&8k5:fհg EN,{]ˡ##B헭' KC3GVn(x9Yルszơ9=3մgSj;|m^ּkF>Y ;g8gp fgs1Yc &a9M]856{wY-ߓ1 /e֓c:Z[c3N9cՇ9BT y"c)uAɕIhWr~5$Pix`,(U2xi23ŃOi0 P6ar39-e&(i;m _7yRS&ԟ/z DD%cyn·ԏQ x`tbBQi@1F: lp6HF'gq<H1Xe;Nl9) l r~aKqZ/n{u 7F-&~;ki7٥0ȣc|`HE@†کO %{BF4DD'dGFӾQ=>;Z!)wW%d5pn5 _f;Pzϗ#Anڂ7% } 9=/C$e/̯J ,Tdhdl14EQ9i>9 uTq~L@[!AҞ/ su[Tethe>h|J{qH#`GzB3/%4u}fzoBNr c=HgX$b>,o;aK96y(%f} GAZ,G|~^Z/Ͽׄh-R+)&[)Q[M*Ffd j0DNxk92wfwɻ6X;J?vyVv?4~8β eU4~Qk %I.P!`V0/2" ef9Oٯ"88fa r>0G6Hc.cǾ" #$$5pFBb%#11)DXҠu0IX99H#]c+"Uu.UlZt{gM%+(uDB L _7aX\7^QRqXt߈ny ,ʀY,C212|U=$OufKpk`k.PHx3jg>OLj9+j$>gu⾂E#v]n|)iEYfz^n<)u ֣N[Y{&UO*W|5`< qN'"@:J=H16y0%Dl>1;ǘ=Z.*7 C ϧR1G EέRȰ, Ƹ#aREb?D h$]x|ݨzIJ~=q2Ytj{oՇt|FR|;hCu#+ϤTJYfq \r^Vv<ܛ0WMՅ0h̄%&V2Wj})(%JJLY נp䤚ˉ5?}SL #z!UT# be}0z)#"b1h#2&"Y* _Z 'd덖M a>l7r4w;k{o{B^^ӘG{,6ʌ ;6`#*$;A 0Atĝ2 }2kl 7*y6ʞ(T#sɴ9$4IK,% , %.C2Ijy #Ou\=2tsp".m:v7aVWޚ :z bNzrȠw+Á= ?v}C}0' 1 #yþ՞| l?2pcl/lmɸ8F/^=.j.8s;glmQg--p|ZnNs7'(wcs7o@TQs7ƬGncֲr7܍1wc̚VU܍1+[cƘ1ncƘ1nc73fu#r$& zUYoWʴ~++DB7l;ĸD PkL9o&snr *`#mu`y`VX2bi#ag2P-?i}-SAku]=@1$Xj)K#aH ET{V) Ş;1(h!D&N{$`[p) $ݍZFe)A@kY}giհ7IZZ洒+h ۤȃ}ٴvϻ8i+>_]A˦[mLjX30 (YjrR $EG+RdPr#<* LFAwpK.R]]`HitL*"Q* qXGX Xc&( ܫP\_Maf fmCVB** y9#AIͧ\+F5Ne9( +TGSF *((!#Jc$R+H.p"&,Kw]𙑺e},pȸВ.`n-Ǽ HOS$   F ]A3JΔ"y`>Ʌ"~[Os VWE7kɖmVNWcPF]a*_܄Š*.g42 S0#T-?swj%A..[9מȴIۢ\-6"O#x9 D$y 6m&UX4 1LZnKI$Lo |Z3S0exKˏr^bWeU_~)Q07hܥ9;#|)boi"V-.{+yaןWRvWd.-ll66ڥ4w޾ ap`C0ǽ-3:ET$d2adrY%/$D4|aHc>_j0j:F:0`72&PڐG`l2 OVS[iΌ:lݎ껲b㹋 "zfx՞ԎS|L5(/i DZBQQINӅ.[K|:NΖ`?K7|xGLE#$f seu`\:9Nu)˛yPHL(A5UI ,wȝr@\em~F^r m=X"9bed%yD"gΆM«A:|{9/ 8T FFcHK#}NZ#Cm"@1)Di0!X2ޞBj)h c-S`.(-"9e"jeޗ*0;/-AKUFRuR.5)Ͻbܤ=XZuv.[22_vQخ^kϛL[ʍa񔵧rTJB:#C #DjcƆiEXl 'm=._ ˁGZDe(j. f"}B;2@%+W Ȥ㕣r&rmԨ4{CL6.Y#၈_V Su̝ 썉ٌҙ]ljTtsWEٜC:6pb[u~89Vtt$tdfψ8%^_^uUQ 'yNh^a BH^TeEINBHu),/^f`*CV2pg2GV+<ȋv+gC&w[[(GhW`=#aS~zzr=_ƥ~\\?o7~qQMяOֹ æc?/O˲i4pǕ"qEٱ+jp15ÌFl^亙~j֏홿.]&g/7hl.~1gCϖb߅Ei/* 8T{]J[?];o?ܪ+CLOGnDE|bxH01A, jͅIY1O#)B$Mv1~l숪tׅ ./o {] :+>Q69A *#"*d3Г@/LMPD!*P\ّ܋v }./ 'ӻiw7ׅ4d~~Q<,Tb;Y̌/vSU<]H3ϲ"c0  ܺC JQ'7PÓ1,Dh8h\*uոC:]˯"})]CozߏǠq]\OQwi8(peÀk7TW.g~Fg GAyYYGMD#Yt2SFr!X_ୠ?w%d) j`aS2& ;Ogޤ4A1+wt' Ѭ u`y9Zg8~k|.$\X)Ufݤ@FuB.sƠeA1񤤓D3N[Kz4:cm)O~/ 5YF¸b-D QGPr2^D5@? F'+P&; s4JZ:[Q=͖G0B0ԛK^h<"i7"8G&>Lh[Rh[(0pOFǩ gB4(E'N:$w#=faVlWp^\+TFWڙԎx)-1@ /6U\NQn^xCC/E4pZRv?UE)HsopĘL8Hxx?V eS gmgwR, ,g oVDȍ=9'#:ĩ:It,EF !„= h|!P#6 HD,1n+p [JW&c)ŧY5&j" () t@ t) PD  N I2WUv93ЮIT<ˈNBBk;7UPf@GX / jnή 9V2 tB,{FRL p "i4hPNPiEOpJyO.\OhI-{Gn}Md( ^,UʻfܩڷγK-i@,yٌfIk DMK3w{jYmCwVdϔ$depnrrLkeP¡@;I\cN\ч}^|Mϝk'w5(;RPƬGP6G;utAT h1N-"ĴONگ#ꎇM0T-zy3uܳvpv&ݥ[Ovh6Det0ƌ~a࿐ [ d`\MWw^d+Fvxyzq-v&wr>o&g7#i~ ߪ6z<ۛoqLM6 w՜2ƍ㞛Va.! 9#|mhE6JFC2rr,ɖankݿ7v%<~dkZ/~׭â'<:,Mp|Znh]uyiX$RsDY#7EcxO2)O9蒹Дdr٨ GkNqx5-57*> "71J(FFuREg !@ԢH~CMPylhQh0$]NXo !Bd%J;\-;%/%|R8()_ףqKmMZt٤``ցNL:9hm 1( `[`RY FL +H:C3*|) #D9LrL{"^d1([#T<8b8S⁹ŅW/i5d6SjO ,Ӝ7no*?J9^WQ՟]L^aH.y?X/4_G7^"RgtMfj0gɸȕIrU(S,%W͚YQixX4w8qÏɦ-:Y5\l.1%>yiTTt#zV:r`at}_}k"3ZpD/s/U?:Esmxg5s\m`x'qJ詒=doǀnozF "taxݛhՋg˟V37GXH_qv*oqE]ӭOW~ ]_LJ dsQNSnt{&Ж_1QF\tqwkSeԞjOq0?`rFB7R'}ߕ?}8wNۓ cA$IXxwO;Jn[ubm+B'+7 Єσ,D) VI6U)^,= KP7[6@JfycVZ蕰c)ǝ$ L+I/!3aw:Lqv5y"P'Ck~kUDHx[ `i$'sցk˱`ȲK^Ã' &`9eHZD0A_kd<7#r7Zc* ": { n"KE}J堘ax *$Rg vtg=0(G:>dKARv)x6H!Fԓg[ꦫ*#2 lr'&S k58&U&$X6TB 9N I1D{ 9>jKvd2snwuq*.h/xT-ʆZY ̵R2U•uRI].WjI\g 1d%B01CGgPJ{ xN&VϖɹYm*/$}:˪ \b3z8Lp>_p⽛c^  +#&!.Q#t0 hT>MiP#T:Q`v(J|}v{(g1 kNk4)%d:gP2dcpS4w9 ޓn}ꚍh?Jdo%Ob;|^ inLho^HqQKop\t{ 8hyUGP G<& eR*:Bg`naDWˎȹHߌ'ph#<ӷ 8K߂Ul}sXb9 >:J8=*ļ񌶘E"#׍u?~q>jÍ%aK5`j)dTk!5׮$vf3ڠL |FfxrA?},AR,ˑ5UIFڡJ:Jh^2ɒRıLGf̵ ?[#c=yOڹ "N#m+t8_m֊ $=qrCӒb%]Yr \Fˎ'#OD&ݻXT6x &ޢS[r@Fn3z0s\gXmsoY__dz+vWg=wEǮ,W,ڴoY^ q/>mr12@YKcsm7Ir-G^,Ǘ$$w>d&0'%ΣV8孳.m"f0i-CLA j|D3DNd(ީMe9Zs%XupѠoMCςq<]n=?C"k^ԅ{5}fz x2k kJKxv0{I96/@hץs;Y5n_ZMt"-5eM-M;/Zbz^k񷻿N9divs%Cxal|YnbMֺXMgvڡ1,6\N?;=ȋI|\ӽ:V~}kzWw8/Ǝ55R+I~%v] L`cx߆q"u2$ccH d$EB %}F F7Fx2\Pml'iu%ȭw}g1 NЅW ]L[<W wkCsR(#\Ks2UN0 ʀ'޷=twgwx㫨ey"st)S,{2xϲv>2ƍ㞛V-ow+/_bqo Xj fKb<ɤ`yiTTt#zV:r&>R>qy课5\nq_-89aTы"9j6K]0y8%TXc@`7aS{~:ŋ0Mg\pųOzI› P,R K25Jސid9'EPq4kxj FW7-u=:˴ᴰ8f٘q^Vq5K_#IIKuGllq k)qE IZwW$ERͻ(Rv -u|UzoVdbnLE]<ŸW~Kzr6tQ5;\]sK~Q%Y. e- %ƛ[] YL Yby LC_4 ,{/>ٹV N :ZesjM,K"drQR DEqT?LEaƒ)A"s|>Ij=LUeuܴwU/~_wuFtNSd{&򓯀ɽt1#X5Խ|Pp»z^pk`@xs7|xLԇ˿^~x+3?KgtA3wWޣߖ4l?M:[Hٶ Jh^mwQI&`rPlY'2Jd(c0Tr%VDmw|a"ʊ}nÉBS0Ŕ;TkgM)j(HF-{tHit nO]Ӵ&C6d@o1 ϏNjk;FX䃤5AR|t%F c]z0qKgfƧĒ1lJ--5ifd;Pe (y0(0I$ 1R`<1k0RZ*)ttyc:7 #$ Fj*e;Q`d)ıH1ph Ow`hl{dsW&jqLo]_ ttJGz% ĿkYQ&s]eK_ͻM+z/k/Tt0bt:9*הfZfZJ[Hu$H)B *^g &uD%ԖhQjcI&zG-6pT#!M|aUq(EYkQjkArvgfm$X!E9o|Z~|f[!rFAn U`hi)1H-]"jV{wܓ=XN\(&d\ ȨIDH"!,Rg2DO S:襤Co(9 Q Ƒ'{ٺddZ-w>}\$hW!pN:WK2!<*|rCrh`-1A Wđk L){9Cp Mm7 7r?Mu$ml50x9;g4־9}#Oع% %uKnӢ$JђdKk,)QO+LVV4񡭚Y1+*ω#,J|{F8x<9[? Z7  :M(rn Ȣ`;X+\$JH!QFRd)Ɨi!Ph}vCgmF{$f%B@`<;_\j7=z4HKC@%Sޗ&2^HJBM zfGdAivߓx=m"i\@xk`+_*6{0}ig0Ez[ DŽAyyI>ˆ^FcLd!Rl DԔ1IQXJ"DS.2&"#Ykz=댭vݵNʭ|wI'eߨI]Ƣ~=?{2̌,k囓䛃omsz2"wS;i?ͲBu͵&eju5㓪%=币94k)-g6pf˘-c1,Ce'f!ؾchy^1-!MWTNmRs]A(NfҶ9d\qUrw7aJSr [୤knXxX0II$ G֏mҸLy3:fWhNZkb:nt>YYg K)m7 dƥ֓7L}Tef;1y k2Qêˊ&x`Uִ*(ԉE_Y|yo7P5lOǼ6N2~|yyI,pKoHb4j&AS#Ekkyn% tO6vO:~]3æUGL0o\p`)$wǮ׭\1E]+q"3 WkL)nn? ^qDƔcWFXɣg#yeczJ3~֥dNg 62(%<`Z( $߄6눙$K +7‭iU"bK(F[ Hcf/h bugoJ.eskn&`lf<W[>с^H9V>5t!q0|>f k|:F,iqD*6ū)Q`\Ks gQpW1VB jk9+H3oKLzKm\ YiP͏z?Eͪ톝?viE̽N>vЫ0iL* BiK4*5AR鄡گ/6.Srh%V/q9~[ҏϗ3eZq߫|H`_uzM5Xks ˀY,Cj\Ncx,׵ EǸ.YMzr=Mo т*.h[#Np2X4қԕIKFMZ:clAgo]^'_V|<[‰fޤJ3S>5(z*;(IrIv|z1QN݌ùO%U~&JM$(#&@$aI"5w 1 "h]jֱ٤J|Ai}=+صۋ5w>@88jDOTdN7A d͢od{JcXh\sl (6M]䍯BhvHD4MDBTKNBx7Eh3 ̺xp1\wwuQ\xMT5tS $@ DbE}0wxIT3"*mN1={j `y<R-7R&~/o70Ah`mÐ v5RaP jLQ8oO@yPbxӹMw*߾$)AZM~Ty{nup5-M~Q@qc4~?Q4 )A\ jM5:tOw{c!C#@yCaoG HHo=_X\-jQZZ@׼dpHknQ_VQۥBB,ZZJL)0"~ ld1x^I?YyXN\(F&!WB'2jXpJቔA d ũ<'S}8'!٦!HHo:ʐ|y#񮌬(r .hL` Ha$r m23^mxkſj,ENHHҺ+xl *8"aͥ&L|-aJ(K)+9%&zqyfΆYt~Bh}3JP0 RדJv67Ql>Z_$uƒe^;rR# LJAB8d$bFM&&7ǑDŽ@v_ײnQoS۷kj0ln5c£egv~&u|JZU4EMw&{Mbɥ_kGsƻɅwdTl6s{fm]9[\o8t}UKD~᷊U#5Y[W #Vn031-V&JKf1uQ[?!W:}ֿ.J٨5$ˊ Z]TJ?.4iʃe(Me{=N%oKusSuTNgO׿8iXʪEe+hqgaq}5jaf#$g\tZ5DCq%|.@3B&DaKL!aqkDB5rdG{Y>Շ143pG/R9'$ V(G{>wiyTe03N3/gpśo>yu>H۽JbUEd |Lzއj͛J?uƩHKe[7l=+6i=;.hz'٢ӭDNNvcuq9f{˗&*:0`ˏ)v|(̈́A1JB$=HCJRIIҔp^,e&2>3f<R% {IϠ'ov].%ul!Z/mR!VW0s㛎oovV1DC BYQn8yL8,>,-6ufK9:Yͤ̊j2Ծ9g.̎1;c.;]W{f))}Ǖ"ўˌit06ZGt#.?G&h7b L"t.24<&o,x )nmxeNt6J -gBFK7*@G^SG]M-rKUS}(4Ê#_J0l=6#HxiYrAb1pY. !gmBÍk3\lC+q#qg~WE8 nni3oq7 *? o'f x O&CKMԬPZ<=ԺU{ΣeYJZMڔx7Z;GCU _}]ߌuL: @(ktEfWq !nhN'<$rHdl`3 ,\ko'Uܹ{ΣG]c!x^''jvC-ک5U _g0>iп [_fZ-VyXAyƵvq=ϵ\Oe][oG+}]ԗꛀH:aU&L I)VE"E %J-NoSUUu]67ZNK;N޷س|Kǭrc-:OҪD[Hsyv>sZ{}%|Xt' -i4?n.Qk/Jz-JzI^ҵt%]{b*ڏT51.rT`jK*wI.%rT]RK*wI.tsPL)VTRJm])R0L+JR0L+ӞG72|B'bu)1N To%)Vdo[)V@9eKRXRpZi[)Vg:[\G+[)Vo[)V@'ӠKD[ҳR+JR@ D@)Vg~o[)Vo[)VoEWTT(槲P)W?vaBOS%WVDTu @D @8eE|ʒvxP5A%Vֽ =(Ul9JhBHԢL@?+/8THFSg-ךROm]T_$с>ipTg$uetvȭ"v-R^kR= {ܲF㞟lGEw []˗ C=]NBmk>V2xg <_vƊn+ttI1z NyHD(@*8?|&'ADjιYRHS ^T0mBZ|GDu 5,jеG?~fk-Hif!^O?\`+&v5DkǙ%%Lῢr7{"10@9!5+S0@h) '9+20$[* qC 4d4 88Τ&Q1xΒj}a9Mgqp7PBipUϳxU64b!WMNsUǾO_Dj -h(# rshKB P +{Wm oWxoIdGsu; Roۢ<~4s6d 5N݅PI`(4F$% wI k 5e !LV>sܺ}y\TD4- 4HͅwdG}\vkYW1aC(rrG a:*e3@t1t.rb :Ύ|՗f "T| 0#8dyI<@)J 2eu2@#>T9"@a.Y@AN?1ce6E#\)R'1ς9jP_C:Z~/zRѓ{U5U>撷Z&"qd+9MR4QU d'אjnN$N/GR3ث"2,eEE.t4<(1b\<R[AXτU;r|݊/zr,|c曗#߁=p65 XDo 4 :Lpa"b0=`PI3 ٗ;XE#$7W&uӤ{ٴlZ 'sl4&HsgSۿv䀒R= SF"\XP:PnW;o3#<&q㎓$P<1 Q $K9$Q$lXOSYֱ*B3@H#ଶ"qeL  5:j:;&^17{?H2>O(Hs$Ej= d}KT0iq11:Bug]i1_lB~ly5V_5ݬαQ`d1#}WA>7N[)gk\ ~N|}c3eFE7!I҆nF$F"Iy_5\IoѬK=#O=^? NLl:jg2)#o:?FY$LFǿFO[iv)8.59hB3Ycp/sz9~]_2̧Xfïvsv`b94MDOPVa"=s%9ܟ$^dI웺}6[K- &oit6#P(OW n}-|Ȍӆ|6Pنܛ@L#EھLggw?ɤC,F+K']\,m#H7h!KP[kj[Wْ6\ٓ̓Wӏ̅~lo%[\5~65bsq=i$7bHaf069p.f4=wћ޵8#7ݣiZ2*p.o^heY!!7J !, kWI sC $ >Or7UY,׳ᵝV5Z/ftϯp^soSnd{&2_ 7QUohJ;*~-7=|o(?y7Woo?zeWg}}Na ԃM@Lૡ|ˆ7ك[pԍ6-$FhjN-sVL#Q jN!$ad9 x4:Oieci /KO㻄]Q wl\2VdS @mmTtaNP:9l:rDtHI!0/)("d)qb$f)%Y $(Iۀ/O (DqŜi80JR@%KA=n#`aJ v],l}sj4o(`\<=hDTFsJO Pύ>"IYMCQxN9" EK W;NAyADrx#xPRGœ˩wshE 5uT>+vd}I94S9NC ٭U%ka1$U֭q/;cro:Voǂޅ|nӚbRÐ8 QJ(ROacGc8", xfn[VƠbw4h&-c&odGW;XMD]] >J5=B5;V}5`C`V:HRF5؊(647g. ,9G+BPXD)I3T&H@xF m}0Ӗh>;6}.JfI)|s|sr|YeΆ4[BWan)Tݴm7Sdce g˜11W4¡0#3ŝ[0O4|`@һ^];(6{]>AlBen3OqZ1IxtUba IY$M'xڿ7훵/qG'x3j{fɖ>o9J0Z[f <%?B.s ԶN0I 1ocƙDO.طw 4?.M;B:O tR׾ݹzw8ۄHAkz|. nE xOXכ֥MLЇ<\OfYsY**'dm&zֆFۚ %C:rL$Oٿ< 殴U<Zx0PjA]aMY\Cb%v|aE8=" ^./:z9}+ז{ z"s(ךY +'^N-߄YZWp8ϟ*^e./{Fr /7 ǻȾ8aZҚ"eWO"ֈԒ(ykK {zw%^ ~4>2|Ǥ9ֿ%j6y_~eel *l@hxyc9 ث? UWv觞Ts0gG_DADGxZG @2muftrBV~:*&l Ff kiL/l,wY6|FJu#m@8FCHnGR)o-I)GOi4i,,rG&+ R'3˜59 BWVb-glާDM/Erf%:Л鑓[ ]a-[kct~nz~ >lڎ9Y][Yqi؈hY2OEٻ=z}:woֽ]f䫛k#y3=7G:hjo])Mn8y<8ao$/nݵXlWo}kx|{["mn.nUstOcX7bʾW{}[ѢUzX_\|9F:;}>Yl~[ e1-vHSӕx0 lO秿~11Ld"gS6:gU< @2"J1RЌ648lUs"3N\SdQ*+^~YE!!*c͉ՖE-`XMS>TSRZRx{볪YRU*op_rUɨ)C tؔ%$z0hI<&3xIs-= |&!U[2V[ݰꡔˏO [mu%.h\Nt8 _' rU[VX!)D AC2%}9]J't,[C0P(eqI+PqF*[ηx={"9Hhrؾ|HTqщ3펌6'GTdU qe0!,g}f/@ zDm4鈙'.n|>O㨆RE0 A8S.v=g?~*/HmBgA!) X ːM6P}ܺm f=j瑹K-\C5&bjxvGMX JK:3He'KIUHe`W4̤WV4!iMܣoѩw9 Fa3zn*j鉩H\m99ZԖյg_xSx:,^[5',d=XRQ6@,(u4KiB8omT&hP`EX'>eE;jKd #VTbǒe<:R(bi4QI.`,ུ $CRx -f6!fStC9E`z igUzգ]Ӣ:Ȯ]gVi=nåk?m>*- $2m@/ o4goTV{Zd;eFuFN_C3H_jjIk|$ssGfQ}oGW٢R8oZ/sͧ퐛_DZ #:. RÝ8sgiXEFhJsz¼sBi=5Wo ί5|7OhXmķ0ojh!t`6 h@H]Fѭ<\2,` .LQ2 VIw7jVxWQEbA佳&+ 9 VČ!қP`?dږ . 1'vRk#9ПX;rvTD2tp\3ΗBObCgnWF^jɺ/^LqCcriJsu{^^?P9imvFd ٪@뜐4g5&bR(˞sqn3}y١@}FOmIj:* J*Rx 1kBJ&*Z*=#5x1NBihIx9p¢1I  fHA;YϪ-gG>k `PG7*]~ ^کiSRBʨQ349'0-Mg Jг@MOAN?1; '@p2F(U>tNq: Aiz݁*zU jדz=iɼr4:{Yi,Xѡ$n#)O:7h! B/8 N/s5SgYfx.4:QL4T1iAӊKm!Ң,D%΄1s[=n=F"we>^ w`hudċK >*+QIJ*oWYH"ÊcMG'jrWC|FЋI9Xvw[fWӒ[i@ğ,`' 6Bi8sR5܀ * {'i((Nۤ,&j?@d,+R6&TeKces YEREgfR; /*+:DP<"'y$8[O*T !sr,Amt^k9;^>wv\ :|oapZ-̐)Rh'L$K&Rۨa[x MBA\hfYl&IQ%sci YuQZR0g5nˏ-Aᨬ{ }>{f]c]?X/V=IWE,d1iK3K>YzRbOOCQ]6)6u\'y4Z1nQgmQ利gt'(팢#r} ̮$3xfiO G IBugMi1el:.&_h0/oZɆnLN?s5#iGi/'ll\ZAH|6A,߽hzD2{`Iuf0?8̈ٸȕʽ򈥓vʈ4=_.NifpF'$N/9faDBgkaU }F'wkN*Ff(|stA1DWh%K0k4{0Xr-Wcv›ŀ\̩d'GdzjoWlqxŴ,o;XLZvNX9*#M8`ƣb'WsLozYykgY5Y1+H^^s\W:!֪ɰl*׸픫ܛ.5 w!´#0$DBS!<2|8 ql;w^;[;Vp3s@/Tt^rpڤ೰Pp%,x#+TiK\cIMpw hn@H$$݀?݀6Yn@A4t8g!(0fQpmh :n ]Xr=+BPXD)I+T&H@xF/j˜˩K`%@Ai( I:&Uf<=8dwdu2;=j0歱hÏG>l݋Mm{>㇛zmʠ _t|) 0ǃ@$Srb=p`2Qs!g\Mļ:zY8)IELG{J>ܕq݀trkA0kӛf9YM77Gk  ceWG`e8V[=[cR7Ovx]"Zpf_/kq c1eq:3cuLqbDMˋi I諾JJ-E*4l āy_Bڶ<ZWf2r$\[ŷM)`yqIY$M&.?4kT#"2Z1ĽaE:}v6lsqtJVvdfQ~LcnΘR& I=EVღzNY4"J}틾ڝԿڬxAMIwp<-8)ҋ{E)("T9%V\fr% HS4*W6 X?9rv L *U6*xxC æUg;7 _r]yB%c kq^5OW񶒷w_M({aZJk+WT"p!(YKT?vX֥ow~Z)Dd vIYYĿGlk'ˉl֥I@ZXh.P፧:c\k\y+`$Sa[l9D#r-,P1,㝻jB.<{jxxz .!X05h]1^-WaGRHtܢA+zMVmӉY70ASoM%翍z; n猔\l_Eg4kk"dzf q _H\W/@Xӯe1?&{ ڥMɳV^HSC[tܠߚj8vŇ?4ϭMpo<UˍΖ ͎i8-~?~;$(H9wX#+cC̦zw*%NDϓ NonռT5ϧ$TV^D:-B{<a1"T'r™*1О؟|9>eCgOM4 >ۿF|/Y Rq O_~݇hx`#s) 2Xm]u `C6\pN ݿBJyb6᪅5/6zV2c&Me iG̏X J)11%cN :^RM RD(ĄAE>W͏ʆƵcu@F1F~ rv<-jK9'76)VG@TidJh+|?h$/'$ 뒌gtAY66JAG -@Ah#~ZbhAlr\ud LY"!(O! `k5I,,3(Ou[5lP[ dܜmOOmM&oxckzl얩[-ԣz6vn]Zƥ5yuս^Zz_>ei$n]5 av7޺nf~\"a:067=nhb?l l1wjuHszӡ_ͳk zo;9AQ+I̼l:p0#[|G:3ٷUm}\mo Nכ)&뻕c\"e"si=Ѽ"'΂*?>9Dx'̛$%Sԃ`j]D HA(Pg8E]2DHGrS$h ?y#8ŸZNѾ rh /.!O,3_3 vKڽ2ޙ2BP/ 'MDnRu *r5pw,8" Mm*%vɗ E)$9僐q'b6D#sKJ=9F<#ZpJ$-<Åsa=76p*)cP IFksNSR\Wb5D ]-H跐)6M\n>]b J"Y?7\<4Cyyg 618ۇj6}a 7b4 eNUWPIΙV2+M?; ²\j*yBmD!dZUdH>:K~AJڣ &N &kHfEy:鉶6S~[Ύgcg}u?F'j=뛭"Vb *pӌ٩ y'N}A5^eYp$Hk*g 8qveQ䴳F"2Dm`u4[mMr=)ZcKnP\l99x%z|k5W_)t'ۭqޕⱀ* q-$9:Z߫e2z_~P$"(E]Xdp*T^k0Io#8 u{44Dgo L,HupqsjtNmT:*8ZlwiyzSP4{׫ψ*x_P֮iQkȮ]ˑ+]ʶCZgBZ3W/o]rD VhDVwwޓ=j`P퍕)Fn~S \UFRQV6 ĥTL$А  JL l"n,Y֗N)'{(ǣK9DoeMjZ}oǣip^jXw>ot!ZGr>j r=zhKB   ":-¼-z{#'kj S/f7?G yC )`% ]) v!6VJxN~q`삒XdόhjU_yJϬ;qkVI@ј\A X D1<}6, rKm{l>LN\!0XC; p`˗n). J*PS\[ !/+)ury9>Zˡi/3KUr͝k}zgbI`(4F$% wI k -e !L6>ܪ}{ٕ!i)Z9K/4i `Ɏ$p!(/-면YW1aC(rrsN a:*e3@t1t.rb /gŖ#| YMMOK;53s)J CJG~`J]n #CCBK""y$%[9QF;!YfwuMDAb ZdR@g.Y_>r94F09^CJB')GS@ Aɩb7:3bIfimT[E(X:yrr1u<%IzCVfNG'cCMճdgz'-sD1i4)brńBJ=#5@hHp9Hk2J c"", q+i|1?=Ձ +"$E B  )A>V\ ƣ̒DvkclҥU3 ٗ;x (րv4dP}:jZbG5ڐ('>䕒MSuls*9Qiv|ώPr*IbHg&V9MwF>31nzţmRKUxt D"zˊ 2LUeKceAȬ",یIh1nGޱ}rE(!I ңErJ:&!qcQvmw EBOΎ[^ݻ,|w' Dd!xR0,} R< m92Ih[N rN&׆5.gMN%rrS: Qf rF#AjpzXYHwgѨg;*#i$骈НAe.J K4_ ]hpP z_CO/t`&zjtd9L'k#XȌE4d2<ːT)-#J;h@yr 2xʮDDf` EK&Sl{쓄ꃣ) 0.&%|Tj~޿c'`>j6%W.I%8'멒G`*?GSvtf  3GigV7xcbN{e36m(yn& rN Пָc^_-'-M}cd)ͧΞfxҟ5Wy~rqmw 0r hxvVފUE-Y;J㷒h]I+]-ۗu6Vb,Mƣ7k/6t몜ݿjmU克XfI^]G̢uWR^?/wӊROY(_I檈c8"y\:AU zZ!>U?]nC}:߫5iʍ"woX$+X2r/UCQlU'Wy\$OxϏN?)p{O?~xG/R(EZ@p<= O?SpR,$2& k^xQI4 rl1+s(ŀl" B4^Y<xBiuު)6`vQ9$23Je6SNt++s)mld,mܥ]ZiJ@څy@+y<.Jr|⋰: BJ`~  ^݅^=;PQq+d䨽BsRQrKn0uMQ8G.<{ݞ gA8$+%+K7%,Rxpڐ"Ss "ۜڔBc36S`N9c.ۧ+Osy7m8^y,+kR] +>kӧ|e),Jks V8-4Yd7[X.Ώc*pI`Y,-B?߾iN_ːo~O>e76~JR?;?O?Bbz%ˏ? Z%a'}XiMD$/'S7U˸ȻpY qYABdp{/'|/&n)'`49;+歙0&]lŇK$8fX}ռX1"_a]x]8\r/R *˔X_ Sh\g厾,,D)!t`13>6w݂u׉DZۇ~E>cF~K]7:{Z<quۃ0dY7=$<[[GMQ=l2}7=-՜pZ4rrv4-riKem3{só9_zxa7扚*~BU GԮ ,i.d\}9-z)wHRn#Km|<'qgsH_".4l . g2=-ɭѻga9y`:^-Yl\|fxXr[i`Q1c%kސS8{ZnM[ʺхu[_Z_K̭th߯xA}weEcގ'FHWW:6^Vu~c녕 O\W2:WU4LRP}Wty] ~UaGr+g.kvʗ;`tw A3h&zLr=@+UZӠT778!7]oĊp!T^ߜ4yo[&yuz7}POX+ zMx-MoX/ hn?OfPEڹKEpok+j&?Y>ety6XjZ?,.gyNh?VMi״)6ꀬ#yGQcu=A&Wҭdi ǛՃpz a>hֲůR*(ƭ2Brr*2T}oԺ!ݗ$ U}Z #s@5d8l0$o.\Q 2(m zH#tυ|Hf5JNqpDg<$ˇTvbR]b!Fc[ƒlAz4yͿ S;ʆ5`#

P>B(k)d$e@;z:&oC@FL3q!2jxrI'#n6Q%Y֨4r!hNe#s4Pqj h#::*p0",0 E*CA2H'ϲ*2ocI;.;B:b:K4Ydbdf)4ͫFjWrV-=l'SX@KMw afs'D|^3M%T`=ҹdB~a-Awy-:B4weWxL;O'esML0,ԵW@7]px3 L[ v( ,E0U]i Igy.j HGȪas@n05Œ!f{EmF(B0wzB #RT"( wB)Wv Z276I+dO YW*ϕ+Eӛ)طyMeeȅ =E 9RŢ4򨖌50 TRXx L.1#dnІE:zV19EKkLT?RkO 5rAP/m(A I 2t$ 1X?FrakP*c>/?LrtqlY8ua{VIYY? XϣKnSj*qq2ivtZ:)S7ú֐ZU78?0޺/6 پne\2ގd$5okޏfתdzj0PnH/|*dV.!5ƿ|]c40 hz(^;hRJI b !UؘqWM\k]5iO gtW[ǂEv Rm__ۨ'I `N%]:%s\ .wGDL̈W5i]TRkc/sWEq08_:=i -٦@(;NדUt Q]BᏖ?N;,jԍYwb~ʌ|~Kj"c޿:CNt5o\3kgnYTϟ~6<+ær7V)1t,Sι`;+ h#oy.~v|$`p:}{;| wH{~ؽXM.OOGi_\m %{|w/KeIYRƵ-n1ֵfq7=>:F!VSV ?6[ ,w>k6l=ǵ-ÿ>eaize}k{c\%>C!V/͖V صx]OC0ׯmpxC>ׂX gm:k/|'q/~;B޾x.~|۸d:/tvjwz=\û[/F_1&x_zzz>_Y{{ S4}*&]wZKۨT[7:\ջtNf _8ŀÕJ[vys@#o][.]?ƕ~Gv_3 C9Nefx͉M/忷^mܚ=O5n݃oػ5kvkO3YIHgG-ƥR *=);\,? 3?Kon._j;w|f5*&,&.E}/.z."^yM%iVֆE\"s{?+ ~ jLY"8tV^nRJCeXmg.%S}jqr.U{ly;oYus0_X]õ%{p=u 狳'Oƛ@'Ld"j|`)wiL)x0`mV ^27! `oǖv{q9÷\]U3ŖW];l[b\{f EW߫"k+5RXh́(I%pTg*T\ԎB8`@88{(E|`X_(|k"-1Y'yAN|:ͿtĶA=t*LCBL(hnS9*[᳭#_EyS^ɺ=`tW㵇Ҏzm5V ᪪6i29ɹ,n (%LWZJn9WlCԨn$E6F661zO7dididiiima`ßkճB/[ilf.,lzZveYՠKJդ~̟̏4/ZTwک|u]m~%-h^گgCkG4K3KU4~ z[:^PڻnWIYfEFU.ZVUs7ZzשּׁȪwǥnrߟ N2:([?d4u-V~FLeQ0]4&yٛĉ_ɧdX٫qr8)<- cҵ<iudG~wZRUzݍxGxz/aN d8u>I ;K^$S1;2qשlp)T57لiWͅRv1^kVMYe4BE"4'Xiə>8ך )uH҇Y:oIYR_xhTE5A%UCx6~juYר`v/~rq{Q(zl4Zp!$zQ&xÕrI*kM'6Dth.*`/  F8˳JWp(Ez5vMr֕Z+N9})Z:]y_eMEkc m_ iƌ HzI#^PnnBXY4ݲ iѥ>23pj)m Ձ 2 *?aDI\H"U` b$S :C2['"ՠME-atgO t=6 :RFR{L{L.SiG'SiXrW7IGY V_шM"aDbarBjB7VaCRFOnspe`JEIX$P qC 48 h8H gRXfA_Q1xΒj}SY:;ÃH@-iڽ&B7|߻֎T9ϵ} ؗobZGv>ʈ\kEfD[j m$ Czť5[Dx INYR50ed;Yoṕ1i7(! oFs}BѰ$+ET8.JdrewgWy> ή#: ]-̸َxUs Ͻ,']h$iLe@ʢP(J洉dB cE0V*N#g~CǍ3ѾS\Z%ڔ諹HHNYvo|9SdsR,S!Aу:UD4PN'5~Txw_mo]TZ&"q#:MR4QUڠ' PkH =7("c7ZOlZu)ZvzgAx309+J!FWk<8fZP.UZIJLQ mX׭"|Xd[ᣅoZz54hlBXB.(8tR0UΊD4j$aA񘂱jISՃHG1V:Xa`1z;Ȇ8T>_S9d$&DFh<糿J@I)TTrJ #{DlPѷ'N}րCIǫHı;N@gĀn:gk,各GWZtKc=!LfYaU:υf2FiYmE"0  P"gúS>?P3}/{WKY[4O{C֒(h,"7y (`nCʝ CH?uB9438ayϔL)2j 6 Q(@?p=UkBƽ8::Zn:FXguw2.Eȯ-jWn8x9sFn 6h!džo}_i:SjgPcQ@5MPm4!lsE14TV}F T8GRCpLN 'DC) 4ёűOTG?WUmͰ7HA?66ݫV7~mWN|jоaIOUz_1~D}4:AY% n~=lk4WOsJw|d\B[w9Z;0hy!A.gTx?Bݐ+%9ܟ ^`v,vL56[K-&oit6#P(<qr9YPcwX{ 1lA7C/.V23Ȭ1?-M1wq~ghfKPZ3knH.ZK+ nȔkg:ϣaoZ^aY#M7oe3J, ♮MÖOi3G ]`O6z0g=\8#Y'\6d8_z6,UIJ !, W|&8Nd0O%=LojetNP_ /)[PCUZ#BiN瘔cN3 T=ĐZU6M"Pe<[Iru9<^Ƃޅ5 Ť1!pdPb#C;"+@v?5.59!F/Duo99mPY|YB+>%,x#+4iK\cI]pw hn@$t/݀f'C7++ ( ,E " A 3YKѲJgm=7QJ o& O5ZkB(蘒TH1réy-t4iʌOLKr9fM]._ߟ˴cL `Bup-}zwꑇ1S?ΗJQA tM H xmzvfwK7[ҷ7Oːel_TUVlE>_WG~ 5?&0 h6/0ǃ@$Szd,1|Umm"f)~jpy=-] )_m{8v'v{2ȝOVwɸa/||̴Ӧ{ > &_KGFnRnv r |# YR@n8ެ4nF.ejm^PWi6>ߍd.F0<`y֍. V#{yLX&j^5LKXzSUdS'Ұ?Bֶ2=̠aIͅMHn:qUJ:^Ҽn4옦#ǯWP46'F;svV:a~n,bmxIvnh2E^AQ> kncf;cJD[dر+&-;e#c(N-Ćj3P^4ov''b@.W Q"B˘SZbqe!':Yɀ4Eí*yecф)YbOWBU >0l:z]usƥ sBN>I?T=@3Mp.f^:sb{@ĎY뇞ͨXͯ;xp+/׿z;% ڗPyf"W}o?,jNrGCb0jM3g>Z\GWC(@zeoO KD厸XSg> iøRk$ss2u&WSq3Z{T{ k%l89uU{2LǮ24+-)'`#Ip@bGC@dx]`/>egq{8ԈEY  Uuw}<\JJ*Q)+-xT  ?uDĻ{u4,|C]DRA50AJ~9ptn̮U<?Rp442U)! Ҩ z_U:0ƽ<=zԄ0>dz@sUL< 5y4/J{{ts=nJ* @w|η' EV%7 Œ9K˭.#:F l6<`t҃J{[Ƣc\8m*ݹ>ӮV}zukd#3cq(d[G/GyQS-K{Hirhlww0E9,6Lf "'$J& 5Ҡjۓ?Q~( J9\_/?;,^U?| &jPo4ze30i"NY-gFz1ilZ*N>'y\QZ,t/Z KLB'0faH{fV$Le* /~$z\bh)돃` oQ_"nWY 7QK-9+lxЄ^."Op^,c̱r2dž_ ηJ{ޙ_{yA6=R- 6>|&.T (MVkZmSVrZ3{V5JeV%Y'3n4k@5b]8^BFE#t_B_X`؊¹'eh9kRy0KȫC^zг`:)|Gs9zռ{ž*Mtd5mc/P,;2w1s-SU:& HMYݲճM>aUa{Ek^.=FDkl|(/{16og>3`jcpB63 cu)6;]$Vi0Fc?>B)0Cc~2hzQvߨLlV~X-lb$+n^f2J!x@!z407֣V]2؃§ SJ#|D0ǏF]%rz,*Q+wއMT6k:uƁ J dDhK<`,:()w(Ҁ́{cQ$ ḑiN [H AI@LATF/߀px1J>%<UWaLLۋZs:x |~𶒯Yƾ=X/$vJ%l_Fm66h>2bi#hqIS*Ź}FvЩjNN979+.YOyy0`2~8SnmMJ_1# raf֑x7Jj,ENw4RDrNh0e;1(h8iT$\.%6,A>iuDJt(K)5a=54Zik>ߙl3Q9L\ f}_P~\;j{9j{K FɺD]2SkF$fII"6 hL }01R-ID@ ^J׿LLYx3S")9IE$s2JEwZI%HDJU' L,/ĭ+Rd³OWXWa {<]S РER ^{#a( {O{X{nZiA *((!#Jc$R+H.pf]^:Mʿ1.о [b1o"!j:SAk0@x ,j/JupY p/afppOWȱ9߬NNIAcÎi7?T!Uϫv k}!6 \*؜6|T I9rck6ZEZNDӦuRUFi}p}L/%00.<;USڈc=.?fj\S(us\Mq 1|f;#mavG/Ll?Կ59NR<-;74uHn塍`J{5Z}qcLӷ~RklIhu!sn>;=mpr5 lraChQOՙ֞o|Q7dq7UMnև0au7uUZ{%^A._zetY)>]TȬŪ:MWq&jL]q`Ͽ G87I-CUFTM7ff)}_b_.>R8(Rk]HL &|vMm*$c$]5ޠz\w?:yӫ͋۟>x&ɟO>`'յ$h&&׭Px nu]/SxWM@[4qS *bakVqC0@rB{Ĉam}t8"D:hKotc5;≍j̃.k[۰mzssNZEĚE8$~.];wt`@Ē1lJ--5iv![YPeaQeo8bSAD"#EZ0M cf9 Xz+::)s(joFI*N/Tj4vRcib0%x;}Űs_vzA#NqOwAup( F5?DQe݉d/);>h0iT8KqK}Ds HNh9R3JC:R RYU4.Ť,HR1{>kјI&r4nGTn,Wz˘Y5"JaoH" 0&ARr2@{!K=XZ,  ڇ(> ҏ"XGs?cJBv-x_JE %W6JЮt;ؼX,Qd ҂\Z8Sc89yCZǖ^ܧ^nō1RϽ1l@ާkqP.݌˯7ȷSwq >zM'ٵs->V̈́CЛvɍ[j9s\#|oh ۃʆA Ik%)X:ȹR[y1!Nu8,l;.:;q3oB#ٛKBH7 SUPN Կ$aQ6Lrs`x5Xp45vAd!tSW7  ontf[T $g ՈW K5 8R=bFn>شB=6pT+#;xaUڣ=Zf=j@W!-"Tn,QVQۥBB,ZZJL)R!-W[C, 5I˅BX9mRR +!'2jXpJቔA :ÒE*L<1a^JJ8PCXYh6t9w@'wv4ʆY̔su/ 1 ,?ͭn R"DA SrkI*4v/}#vo^|({ӂo7I}[wRa0,R"JF(.ȄU^ <ڧTh\JC)>‚ I#pGu#+7F+9*Fļ!j$H11kJ'0i2ZA۠FH"I`vt}M+rCCwؗ^fdm.o[\{4Vz66ŵ -aH-TBA,J[\ h%vVs Knw[N~{#,@`b0;t b/΂avEA^:pJ@t%#D^N% RjDs`4KB lhڱuFΊr{Rv)`Tx+2 8:gRb썗HcE,DCIA={N>-WD3W. .Qs S T!ZA-7ӡlٻF$W=ZyD^6v1SR",`FV%R-I (Ũ/*}Zd,GE,F4IDF TiʓouCQ8Jy5J/ xFF++JFxpiAVޛBAX9΄k[9ʖ[ǭxCQbKdW壅]zZz54h֬Do< 4 6Lpa"R0,d㛻Cюb+dW`E'm=ig'!U3ꆼ '>dO&HXA?˛iU+;y #qQ %(aY1yY*Ct`Cq __qePK4t8g!(0qgOkCc1pLpg^Z`9Z2_빱1$nQIψGRa>1cJ*S(bsQ`ܓqkV^xUʼY}p+m8/\j@=,gr{&S=QM(GT8 ,q~ zE,}E7X '\ƟGrb-ex96Arh|Be*RC| J1n{9!v^,bG^֥g:=4?Vy/ݚ8+?5.Jh2 nP3}g3I_  %p) bTmAL&ʌdXǜ%=&b.wѣ]RZB=Z+g|nuikƭ~qsuy鬖jyi7CPS¤lkU9,כԛg+Mo\ ?$~3Sws<_l:>ϧ{lI9>_}h}%߽f>b?|dYDNXRRuYˋӒ%Ub[:RKK^@ ! %0.t7ܑzGCBG..P`EHʐTx% g~f >qNnώFs.y+-NŞ3,]椅YĒ F_kNcf۝1j2us=݊/m6z픥Jsu;׮kmItzyT;P^ 'ϧ'|o@/x@! SPD`s[K2ιDQ'4K<hU2}2U *Qjm)һg MU{Lo\RA_^A DŠ7p}~YO'wͽW_L({aZJk+WT"p!bɞ7ʞPZ%+yhiRIDK풲>3܄3N/X$q,3jH]Ou42QǸָ\HYFYFb<&$C ߱lfFQëᾐkެ4ubiR{U{Ě=5Obz'8U8_ %T'Ly g#~#UUrBJ(=UjOUxGqQ"$9僐q'b6zq.)'F<#ZpJ$g]- + jff"r5ӟz<\$=8ż.\Qը4 ҄2*+ch$L+{%5#{C²\j;3)j'TB_*Vz'y/*Y m.ʆ?5C Ӌ&E [҂ F PGak"pР WOhW`oHN;k$2*Q&j#jHRg:EKbLx-*NYOWǓx?v[+%6EcQ+zrO@-!>m_O六PҩڪJ4Re9#F+Yr[y46k1PIw˃虋 ]wKsv[ibuy}gmHС1,6ߓm6msZ QWcLve(}^ 45VSvޛk\|ЈE!3ڔ 5x l!p-1l%:X{Np|yQ(O?M?^JDx'̛$%8Sԃ`j]D H@qBDuF 0b"E4<b\pJ'Sh_AXL*'MNBJ['^Uﳜo{/Qu(CvX J)11%t̻&.WJƣ0g\ jZ6FmQgV$oITts,!YV%03/4u:3m]U mA4-ddЊ<*Ě,2%ǘF$R!@<QAYQ?PEaD="XHhFd$!(\p^C.G#VQ3~övkkQk3(E8ڣxTwdc@=i)*H6)H0ztVw+-@ Luy*")BWBS9L@̜}6.šoЖMfix k0IL~/"lsJJMd3BDN HK|78NE?#.O |aL#+Xqu4pu,p*KTWoZ&1 _r݇_~~}2H W/@Tutl>]>%dgQ5F#BVMR=d9dJzȚTG7JUfqUV<\e)MWԆ^#Tp{6gJ\ zpaTW;zɕ~GQ \c\5P  / D ~F7 &i;4atm?8=䝤=|z/^ %rv_ގ4w@GB,ޘ/(. |4>xvUU}Btr3L[K@ԙrBhw#|7.on.fԺWbR)w~':-UEV'2J@fVDR͌9"sG1B @iAC( ocPȝPvxoso.-cT8"p,>55KLo\9s4peX*K͡UR\MFx7o?_κŹ,hy4WlfK_'Wg'TClQEj*xm|x}rMޟ0"(Wܢݪݤf& )$'L0ML0aQWXb*J$$~%.qTx~YO'w]W_LB,@P %Q@T}\Sw&h90*jJ"2F,ߌn9{)p-1ee&PKZl@7hd!.--Xi#1H *& .rYY,81xu.`mlťrҤ;,@wETΓ>mrUڤ2mr *PѳپBF_!W+d2 ).YZqh88#@TI2 ȡQ5Dϡ{p>T0w4BBǝLM9#f3DҲMXsa=76p*)cP I_FksN?"(W&,YF3wz#r(KsG?p~.ͤ(N1-j*<0jT… iBESp14TsJϞ~@I!8`io\fvd 60(:ؒGܟbGDrҋVSd]UXU")i,LFȸ,OA?觏`Q(%, gݽObOGQ"PJgYr }:F^]J39[^9? >6W_ b4#imfwrw®Z[ɇiٞ _B<ޭuiI3Q|VBqHe'KI:Ia,@c>ʪ9@2hNGeJ 1~;C+X*Ffe!76V&ő[/,or\UxȖc ·l~bHQ1<ƝnPwrO;0b KƇRl2'o Jt06SwV#gKqvW`5Vה>m ϶ ]vw5kIIZ [y\PZ/ir:]Yoɜl; _C}o|E7ZbotSzy~}[jnCg?ŷOd 9"Ϧ>N!RPk*Auq8ov%n8q*%0nJ$]}$xnDea{>ߓ y5B6SyM.{\9KW2)YLJR~ I'_^7;g0CHqJAthL)ɸ  c2Z$u9!. wo4< $7VH-7A`-X<9EO-،h ]_+sYɭ,v,}#b┴"|]pTS>=p9P$m39;ڦAҘldX1&30S#BWՌlN)JTR8̦H䬼ds|&%U[3V#gf\RDV]u E T:A\$Mny7Yş6hY =r^X3"ET(-9fHZpu60@+) ҃YM*-C|L (jS F%+\|P39bfBZla8H7˃ڵcW*kmiTbY"hAEL$(9dE&Ki|xcdڪ0kc$sQLȐ+DK&!Fh8b HdTsdYFFRRt?X?vՈFԝF4b3VيˆC )DPh \1+rK^mUIZ'i/iZ@(*?g.{}Wԭ'iLE$S3 {7Vq%cX&TM`@@X+pTAWQr5jJUߙtԊ(^Du1DX0!y\s(fL%/@Ĭ0!rMU F*yF ›. ǘ Ts +9!0kdEЃ]h ݈qeBh?6lOv_-/GqU>CI+%/wT6EG8SA`0a2dЁ9!]LɁDBIatzNΎ53z7ZU4Q y2 u r#HRQ3^g$V ;Ɖ*Q\Z41 ъAD&drV-嬕?1kuxQcVq NKRF#@DÜ3*DsRr|^Xx`D-*Iu"PS>=OyEIGg95F[kDDǜLNs dAz݁*;U jYg'uvBx;X2dO"K_86q 8?iPGPY:QEqqryq(A.xb Yfˢ f6N;GmTFL%)0ڦ5ʕp9*Y ĘGh!TJZLD5Uxegm} &2 2q9br+,FU%tF -̀4ȝðx,ATV]CT-:A/eclu 5i&LP&%ea n[Z kojϳ;S#!*.6.i;PjvNj˖ʾݚӬuoڼxrV>y>x}90M%[sYtm^`8rj7~=+ت\9G:_5X= dw$ v,Xppv5_d ~$W:4//v4 ]VTbժ=Y)MJxd%\PQ,_I^QK4myizZKԵJ8J4WX\*2M!-+ͿB-7/05S5$ N?>?>o({}N7gD} {^߃/˛W C9wmt{_o@"/=rB $"B$rցYX7+WEfDIr<&r"# 摖30PFD0A_5rcf^] @c7)t(HΉJ!@;?ez:$Sg Ջxtr@3R3^LM h.eƓv䌋Nd !Hb9ZFɂ}uUf&"B'DaS {TZm>;g.;ʝPcxyKbH 1m` bIL¬6DӦ>jG{Y{xsw-A[ /rgP|wf1^8)sy|FCh\l> @iFr<@M \Vrocpgpt=xo-]L[|;u@ /pv:.YbȨ,K:`cr@K礍Zr lS{7K<28qϳ hbp$Lr:pK6j]B: 4>-2B>W܌9]Nݬp3c15ڵ&+lWpͱCGO<:J}:%&_i̜ѥJR*f28hѤECd5&ON(;Hݡ6tCXGq'ư4o٭+.ݷ-V~+b " [L =z3]}4C`ڀnlڭTO$R蟔<肫P~8S!viZf(_x44y'̴}G<#'!ǟI{+|G`K GfNH.,eIQh:[9k[,PkƋS00ZHB=-.HPEϖImt=]Ų|7KGfz^/{?M?jz_Ξވ9`O{ 3;gZudI$;AAH;v"99'YA=ٱ)KrjvJUQn:7?wS\d=yz%?.u.]_#0@Yey=*eMrCU1& B: !V΀G>hߋΗp+)lv#G|əqv$G"V;u8Cx÷ɏ/*բ=r4`fH/?l-DPXT,O֑ҔCl&|hɵ 򧢉X2/NA߹c5.3/UFL@67BJM g5:"t[BӒݻMi8Pow3d牁޺~` 7?l/س1KРH A Tţ^ P9rJ~pLk.;)Z\ͬ@gtCTZ୊'|;#&Pw~#>$Ɨ?no1t3lI!S+ڡ2+/WG>s! s}o&$9ô}Yz~3sw?E. 攄 U(1u1U :Y]t+(Tֵ *{Upz2SFhĤQ+lC/ȑ 9-.$VrM>rNZ:j,oTr;!9Shgw=CC[7N(섗z~=0LMyxaql0W@2#1'gM#P`0+?{HzR+wJN"ӢwP#nAw}P>KZ3|~ M7ܘՁ W6haO6]n9cwKbZ;I]e~Ozv˖g4t2ÓOEuqVŋvGw$Scnz1@<[IG 1YӀ8"B dGG*zX^c=!炵֩dC1F.H89PYc¯KO"$BRO%6TN6Skc/SqTCZ*f=Gb bP CO1_(<9Ow_>H#4$?)$Ψ@))Ԕ6RfbAHKEޭ)JMpBmd51|զU8&4rvLtc{Hr=L^G]26qrQoN'p-"9s៙jBg9_l[N؅w?9[U"7.m.V7|mW__ kuAa~tG_dn=Ԗg^9r=\-_.6l^v➋o9"lRJpWAnOTzAWƇ|s5ߠf6JtGSx; 湋p>@}Ѐ- C ZDKi#e|v}5?*$ %ljó֍ ~'tGήp5 ==9^reת'ѓ_|S6FAՅNQhA!c*EP%cb)E]KډSԐdgOu.o<kPd9y(wNv[v-le~?zV4] C`ZPj M['k8s] R\UK@) `FFG2הT/qD\ =%S&빆 S}aZ[UTm9#frX2s,1|cRkOQfc}}U/Cz#v^D:`MxH5W[^d}RȢ1J%)wl!SQnCA@ 09d}   Ws*{Tk9#vry2Qus1jvgDV!h s li FZ6P6:.XْNTj.(t! 2(,:%pQANciQW* .>鶜pFꗃ!Z`<>sDcD#Fٙ oɛCȁXh:馒m$6fKމ(W] J9@ ^J3S&,aS(gaҀn17rG5t1XrnÝ8EJxyf5q*R#y/`|/_ţU7eyCI4uhsb ~,{WO ,|:ʪSH01QW+))bs%&@LFm>kwr{;9[}e]ka֑VК֓&?`T!R, @V,`5`+0=U._› #TXTk}ţ%JbSVc8|>%&NOI72~918?}OovOxȡs+ a+:xqZ6(|@!+eo\0F |!_,fݽpo-c0xOPzqFoo9/srec#E2X 5 p]ǦVGsrBapBBaI(gf~67#.C䔁Urc "RSk1QV\ʺ5D)|v=nm:t59jCB*RC;YL&sn١5Qb E4 ǭ}r])Qmy=e7136z$SYؠM^ڭ`O˗O6B\ < hDaNT&qa"\aq꜂!Ue fm P(O+S_YN9˟# qb2޵#?{( w9';`05 >me+qWˊ$Rۖ2θMɪb$ lxJG^d)$Q>. {3+QkܺhU]$W";I9)*EbtXUjFlNktםn NLm.tGE&Gsi4:/rWǿ@f~] %?@.?B_ VZP & dr3?}.qP`.b16~3uߡwNd'q詀}qXvSp/]v8ur?vM9~7WK-&o;l D鸜\%`Fnq1fAQTL 1bhf+[D5mG9p-ӎ(5q}(N6<ǷjŖN;}.ْVR[S|x.!8)1kzl-{?_*r7n!Z|eK t:Pf3 m~ Q`Nz8ks뜑+[ed}Nju׿. PQe2mr*¢W P O;q")gү|SG}]_^Z)̆کuZOױo'{ձF>M*u2BKV| e GT. 5UJDIbW.|OO?1o?ϧ:>=™?ἏAV`@mMo[|?hz=c>whi"7ABm2q^ m4rM99qr0 A얆+7S4h4F+l'(n6)j\H&QlJ Nڶ:oSKZE[iB|LBE.[/GA˻1p#rtXuXR?>p*ŤDMu ft= k{ ǿXԱ߮$ ۨiؔq{?o|W˙WxyCΗꗕ!TP[]ic`,Do?Η8=m/CѿHkաg`^z!R\D"2J'FaRu@} K47{?ՓdJ+vd}6;4S9^NV _%ka1$U-Ҹ9XvXڻ]v aZ`YLc2G&!J y@n+#<2|XA=Zb ښZ,7/tz!qdi&`A)/8&*+"@*պ t+K'AQqD(\iA@+/8THZ,[ s?ZTHINYe2qOu?NT4"qT΁47r Mn>űEG/! e=w9x(Gn ZԕEV拺~f +Z HAĒңC4s9%c|;{(j{q6?nE 7W\^oPXyݯq+=.ZAVM}=7.z :C67Býolgt1gDHZ0fMjtUO=Xjn.-z!He1 NQfai8.qPB'P,x#K5/W:XItP(d,V =iMϞʤ`C`V:HR\ \fyi%hwqlscc Iw& eQ9?"<5p1r2-j2MѨbev"g^ݓ;֘8lj /-y裉14}_zܑ$|5G݋/5z jKGcqY D!>-y”gJ!}vѶ^aW6_PLD4 vS@hQF$zTjVj!h)g\7~'x1q] ˃2L0ř^R[i6Fjz[-JG㖃INʭwKhaCǤyůE_ 0AS!>ZHfPfCEB|d0\bPI7Z0w-zޒb/4FLB,K&GmA`4b-3ƀ%:yRӀҨb >P1@#٥jR NΤ Q娀B@c9PE($͓g M FqxR/|142Gwvw?X0#l/5ij }[(}G߹LOZd"(52dWZ2I|êkcoxSj|qO.8LHkܻK<:+p/$˔׽Ei9HY޽mv,S2O \ :_Ƣ >1I)]7 1bƥgi,{*2a;1 k6ê vMC1nѼ,#&W/7ŗ㑛,sv~ FSGD,!9"!!^J;8ict88ܱsM525p@ϮP@! SPD`s[K2ι,DQ'4K<U s:s42tVX.xc\Jwϥa\jg)NWZ(;]is}.sUʾsn7\ ´ 6֖/-0(E B$^zW6We ejon٩UQkT0%e}dg g)-n߆\]w \Ã" Ņ f)D.:H6w+lN(0+ҫH&'t&עW3z5S)IW8z@}Qgq2xYu8jyu8*얺ҏPWUWO#gjn=…<)Ƭ\ӎsj^wۓ C%PC‹'wŠ?.NZR{!}CtfomJ'3uU?I_  Ţ^ Vpc5a'2ԏʚK-@ Ly.#:\ǐX7QN(ceQ2E Mfl,Kw d?~8N2sW5462{<P[ͬФ&fvβHR gsSQo@W`p&ׂ3v g*%mL+RW`C_Br y5*Sy۫L`zꊃO.}ql:̔??2'0P@rP45S"R+ܦɐBBF%"ǜ‡'-80%tJHLҙ(l3EC g<_MUΡ]d҆Z0l_Yư<D:anZ^k:8 >ؑ[Vt]ew ewmG*lm7 ^Q?{z?{WFr~ F > %AF)Z "iHI-g5þ_!Zt1H L]CIG1Ej%@TL{Zre'+lOY ]# aCnnq'(|fXp9 ^>(Y%= 47%UXq+%&FmiGnK4[_*yo9"b:sȑgW U9Grl@uۙe7nzͥE&l:l]?`=w%Z{mYgϘ% bf ‡=c. z{K,Υ- 5D/P7i% RDfqwŲTUJ<ͯ "=a:ޯFFiwMVWZ4Oh{?or=GGfxצuB4W_z5n?.W y eҡdfmZn}b]Y枉Y愉Ol?d?Ŀܛɾppnڑ~9O < ֋tq߳#ݰ:xWg0Q6e8V,lKBnvt?M/~>1DYYLD#y@)"0|JH!X Q({:PoL[\SdgO4<$7Zc+/ךPXm9Y4͔Th:Ɏ6_5:y8-}UmgJE$ٔ%5c΃R!ɀbʨ8L7 ƸdTF+cb.HPJ)阭ѥhrFST\\$jKj/F)/" QCe!TO?d<_?dGqYf݇qh%[b9$֔[>6™2cĭIqͥ3F"c9m21E{MP:( 6 >YMhU)Me]m9%qy4Ru`0I ȒHCI]rKĝYJ*uezţY0EWbՇPxhDTWz|2~܊9{GWuCƒ=^vK!N%ZZJ"DDgyo#D<3!+  9YuF[6 }ղ }"! F> Fր'6龢+N5hNO!>ɗqTuQ&m D55S'm}i4o'g2fze& h]t={!'R:ԋ٢$ɃɈ9J+9t p`ԎpQAF6 J Zi1xOJY)Y:D2s.&$<<,f$pr).u),Y[Y!J:pϭ1&dὦVѺjZ~!IfKB ?9ȳ6zn7_՛O}3r$߸ >zBAlԨ,9e.L|2=—O5܈rT66\x `hQhͅS=ODDPf*'΁q<}. y(o3FI8k#M!*`i(Sn|L.CKCMd*0\f2Spޅr:O/E|v2GaNhO/(W!/f(-JF4Z@.0Y",.# Ҡad']/ɮr*B\њK$+3f&Z !GLEbvo[q7ǥW oZǻSn)*vEoYeYũM?4x 1)湂a Ʉil|2 %4pOٰd!>շ"5d@͊ITdFj "+FbgA `,VG90GcMd >iSFL0dkQj-g?HGQN2UlJ~>s)o:B}: ]}8_.E E{c^)\~?e _M\jjcSVɱAp%t7юB"]P!bf>հJ}"jM0Fg:tF8כiPa#!Ɍ>: t3.CnYw|! XTėgi5|^5tV5V^ umBZ'pVWzuhgk7Ph4}ekSHed@XɐC9|D LAAFq&4=^6ab[!j/aTN /Q*H2+URQhu@ *Ĉ:ey, A ZJ:2#p쐻P;r"LN)y=FǟUO>[d;X}usEp@B\0@LqVF@XpF?AYg*̻DoHr7mNG3[$`โPys o.X76P.b4 lP.̥F5Qm<]*?U]G**V}f`FSVê$Nd+FgY9S^9U1uPFkD^-2{eJ91+ -'$ѣSx > -v$C-#OI sR: wFh)RZw-/?f㻺B3 }*;mQ1”^ mFYB# V |6Z+%bJ:x+-Gvsn=m|v*'H\NL*Hʀ x Qx]OcĺZ[Ɖ7X^.]IZ[*A0&%'`Be>߿6k urvृPZ 1'%Y9֜Gyhx |Kp2b5Y1?>9o\1d5OpZ+RYR*f+9@ɨbZՠVz;P+BrƨHXd Fq7,[MȵDM'+% :V3RӒ2Ni/64: y3aLV#Bkn=;F"e>^mT`` rZsN .H+ I${aEXVr{t^͊b n݀ni[U_kD?Y$&/l)V=dTbtJ(9G&!7>3hw-SRm}O^qt]z΢ nq<,5”&òY2Ecd@Y '*FHz.%.;DTLBZM GmyEk-k:?&W3n_ON;v.)adp`)IU"2KDMAA$0,mcxҥBe49$40A3+4I B$j#t[ptEdG$ZI djVQ,YEݳ6z:zfbC'@X^I;ݖ\f͗g`.Kԓːzzj.tЋlk)c udLgƸqZ* ycH2{FH}TVK?$.g9q0DʶDrΔDO{?*OS)>dQNmgq2xflxFxF˔nu[4t~qU!Aq`_sg񍗨xfO .m+nmDrz&lp_#2"M#(.NBdӖĶ:KK[54M$O@rS KA^vRF!)Qg~sEgvj5RLfӃ"KW; yށ](.Y'ܐU4vO.ə.u|!,]lNz$lnKys4yU^I/WKYnsKP5- p~}&tnh۟VmR7w|4+}]Kz^V1ٚ9zk9ii"ڒ%=-,n/\,^x{5?_&e1֋p y4<;_-؛ ƓOWV#[(uͤs&t5i6gbژƓV S,`i2]zv3b8ksVu*ݳ65jݙ ͐,+"d=e(i_ dE]$!VrӊTi(K̓"N EQ0W ~oo/Y(׫{k6FV7,h0 ).:܋ElC|*.8/.$?/?Pÿ?\݇|Kp)aK:>NM[2Kx Otew |?P4S6qsNǎt2+nN0F3 tQc9mlr]6X*`j#Vlbbr5$A6!P&|eU*+`R9D ZU# YgR7,0CAʼ:;=/#޸S B%gk82}j VɗT'ǖGi?CQR}&J1jۖiC&J'Ar2y(R+2G[֨Xv݂PoЫ0e5J&VkPB)g 1ݤq/?sR\j ϛ1=pRQs4BRDF (ٲpO'bQHbpD~/ kOο[ x3ht;0X=vzq'?bieGF?ߟqBxZjE%^Q9/"EШU&z֯z_#gzo7o]:>_Ohы^7ʱ$t93oC^?oWr+,ýnX5_Ow&/|ry<.zQsqy>/{~,ZA\.rQ-2g7Nj&5`%ro@Olg-vEY6t'3a ] *u)&ө*Shl^w2a,ˮ_qg7/cdo:61$W5\k f/"Xؒ!NQ~LԔ{m\6[bTUS+HIP] lTI"jR5l`oBc⭙qgo-y\8잖HԺ}v0Lנgutz\}-,Ӭ@r^;:D[&-?9dbu+Tf`lAS S-eO-e22+rR: @~5 vbSTb"Ⱥت :R(RB][{wƙd9g1;WW՘Z@֙PiU \a[Q%$X #灻)/`|L=6w5MӖyԳn!z֡g|| cBv&6[Ҥ*Cd8\rzQ `5QVǽa/ZoQ1%A]bTܐQJFao[w:)GCD.*VP*|Vik9gK9\>OsAJt϶=Ͳh4?G$ch 謂G6[%?YohˤB Êg5צyb-cZu&$P'a3JLZYţD4IQݦ8^Xsz88_B,M .p+& QgKH'UGpRA$^[dc}D~e`{;9^B7^4,_/6tHٻNk l*  M :eJ#ݭpHN42#]ݔt4^.gޖ f -̍.f{>tgOKQ,>1Iy6m]V j[r콎g2-o`S4pqX^bg nFkė[-uz|t³4:WBNYu-,֨Tassʾ[9rԿ4@*^ !9ֹθWv1-rR*>zKmɈɮ#DS)XA.ؒ3^W7nSX#f.:BdH]풲] *lFC{|HErٸ-*6msۃuccp9O@̈W_$*5)Jq!RZ*5j zFc7X)@ .{E (^jx-QF%Q]AuRH] ը~Ȇע@Y}J<;ׄ`PjԕՔ oԢ;tuղV=QWWOTiᶹj}urUdwώO~Nk_q{^?+3;9?X ey'rN|o{ޡv7NSi /ۍZQӻQY4iTV@˪yq:}kLA>^<>KO$ 'Ot~1UvJ?[>ˮ8?;;_t_9!orKVi[Yo^:dqzv&q;GZ2WEý6/?7RrJ]5t*ZrkM􅌋~%`x$[ر@ߙ=m Z:)tv-7V9#Pb{=?)\VFZO恺\Jp_O]zhSQ\b$EE騫MDA![.W L>ibb´5j[\Ӷ0km1/ƚO/O_d?9mvO>(Ʌk(೬%\w!ZPν//ϳ-u'o9ۜaB [o Rcq4hЪ[mb)H8K۾[MRSiN%XoCȹ5z@8nMqLPf#t;YSV&NݘyTy4.햋?}ӷCOVدOD1B8g 4\l.$a(,ByOю;\$x#_@ksFێZIVANO̐z*#2LJ # 3d!g#0)XhUr69^ lUFmA"k M(] Ĥr ΂8lFR>GRQ?@O0E&Gv%]ɥXT\kY Z$F# j勓Nx_ C,APBhր$Ӝn8]5bQq 8P;~nǛĉ&c#j hdUں곘"15(f f`5l=/@/Ql]+ٚQv)bO zb ] ǣyX+>?$W~G[^\KY|wQbhr1R b(2iaI XFᾆI._tr.i U8x[42T@HgBcT$XM]gVȿʢ6߮>Fc%EzatF, Ϸ&ZcVFgIS02˰3rs-MxB8Ad+G 5ꪔ>*-)׼iA| 2.ڧj1;NERELZRE;JIvߧ({?*OӜ)=_.N_.4Nnf|z,U xv1IcZw\/v[o4x'[7kGjH뇙$]9"|Mipe Xg =woܤQ93ڄ#!~|ݧ?}O#87zKҰ5$8iL¿D?emF_nԍnqwKM^D0^ 5K`axy8a\"˘LA J8VCe9uԙ5ߋ^ S ǘWV7Y@D*xK; (òT(st;倬$=RnD*GHN@mb<) ԃ+e*cb^"c1,7%_ͭVo>+22?&WrjǸlsbf2Q qOLF QnGC`U@[KƠin/8z3yAlh\l:EcWI8om#%SM \-*c1{:fuvh!4 UDL.Ig|rV uD%@XX6ZVَ tmVCGJޓSgBE.򫩛fEo}HCzҜ=;tDH( boh SuFbR(_$ѣH&ě8h-'IIdUs 9W!,\2O E~5d̚3Ԃj`cxv 4 NqaQ.i@UPN7ZHB| 5Ȃt7"&G2Lf+EŭȽ2 1Pr"u&; 9K$.h(OȲRh.KYf*r!C0mȵ nW#giY iΌwFkq͈8jK b`z=ͪa)Vy"cu:2hgw?q$-&H,U.|zJpOhB lk?gA}+ ׎/ a7A&K>8!4Iym]bi͙Yȥw\ Pk)lHel(a'$xI)K2YQ83%9; Pzx?LPf^v2]MZNx_.>>Xhw=,],p4g*' ʤ*G.ua& /Z1N0&HI1O!Zٕ @g +sNzfs6JUFf#4)3䬸s\\+i0L^Me!+`$j @mq*ˈ|<ۭ6uts7e66Lje|_0apC;Eو%R0TRR:ND2G & i&'njRe+<vg$K$Gi u{r]W4m|t~orёhEmwj6mvtWޗV[ LRRYi=%3jmio4a;j(7 4.;u)D_/(ͽӓϤQg^:"DDl\a&.5zY '<`X:X'^S{Jƽ @z $A&bFÔAy0{ ! eyIH4:A$3GW80j)U{R(ZJVƅ/6KmTגTb⹞(˕lϨl}6 YS1mv56M`L`R$#|"7΅^,R䃣o4ޙ˟>{cpEldT\;_->HŵYLK?Zt:31}R٤qHRe!dޱ,%*pe Px;GG}&~%yf{Z|@NjeP¥*(˫[ˬN&mD5gw3{Ӈ7ow{xeF>+ \;Y| hӐ*KO6@j `)dllJ[uաVsB[FkA!75KbYb<.>2ͪ$=I+ s ݮIg|ŻDㅎbģ,$t9G 7ef| '" Q7*:P9&dh69.! yy0sH;9wJrטGn=ՙF6Fd^љMlbonk\G4G#3Q$Ixc7(!Ns|UX9&I%!dQ2Θ|%8K9=4*Z$fuk|2̆09a )pL[; \WƝQs无5붠j,8ͫEޞm@?5Pn=ߡ-Y&ZK4Mؖ> @[»w}y5n %ݚlȍMz+_|0vtLdп ۜxnϬn&+f-&ֵ]v~Mg=/E[wS߁I-Oukў6f݆2magMnRNk:/Fq~.yt mZhgw%צ)E72$ܥ)5[ g Զgg~Ͼ]}PƛD'יD:NC^|l5|HcMi-l]>Gݷ_t0]ϝ]`GL>4~b]ykk+u Ics1:/z6}$NҘmo&{]|m 1P&QF\xkR2 <)I>VBIw*3$$r K0Wz]C[ O@:NYv|-; Q 57r2EbhI"MësT mr3-* {GiT (PW+nEާl Qb'<D"K9AfLdž! ˎK*+˘]ں0V8ѳ} "RB(āhx\$]->s,]pR} 5ovCt|\;uy<܂D"%ųN$f)rh#LD M,O0'f1(4#HB7_u ']A]PS-&10F0L6AF *X4䌒F$)&-Uv"ܗc'P$ifեIt<:9zEYNHbْvd] %%̃$sk."H ]b?Ҵ&MF;&Z'N#όqtFi yCiVbZUzQʋ5W޳ }b"e[v,UD9ӎh`gcdT~*쟕`JÛ08f[]28I9M2OKJs I?&#Wd~&kW8vNҾk|tV;L%KvZțV̩}0'rnWbqp4|YW$~=Iئ;J` ;n|o*,oIicefbKb'7=*w]<䦻. R˗Ncw7$[VLMI% ŵOD!-tJxu4e2)fuYX{QSJ`JS^QvS.m{&В_&tVa;4a~ϗ٨Q5i?f87apOd^ˣ뛣qG7/ 4B`yi0nW?nx %D:HKoA"/]8J`:4)&*Nnՙu"!k# %7˔=UzT$Bipc<3a6Y *ɍN/:ۗL|%l<)ɻXD-w5ɷ{AY9Af%#@t6"DtpK-a:idVsB<0(gYF- ,耳4`]XiC,Gr)h:=CJD))^3~gp뚻WgA:TQ /,JI̠ϭJE>a8C:He}AQ㌗-S.h,!9mU2 % K-UA9|^Ub:$A8fJL X K{H\"EUc)Y5ױ1X-<:ZA IH1Iz?0R&AN:G:6N-ŠA?,ԃk2 >TÁL +P4ɴ/ETh}o_F)kΥz,ZJUx{ؼƩ[[ZOnp5(@bſؿ>IW==\ [o o2D~{fxZϲˍuq'[7w G2aQyM,*vA$V$(&^{$<0-qdzɀoee 2h&&< vGA#oOazϫOIV<0jͻ<4i8~:ͽm>ݺ8Vg9)?txHSL"HyEbNI("(h d8UA]@K/nCϕޡ̣75Ҳ+ɢ\uˌoi >A)t4lt^Y?$=hߦI}zvӫ ? GRSܩ"Dn WPᦇyEg3/H侘ui[2zNݸVHH Юд>"Ii,QICs>V޸r3l!!xY9nyocl}7 j2ӸV(SIn,HfBeIۨT*)s$`9,1RdZӌ\SzVV4ÜL+ε,DaL )A (}LjF4O> ߟ-0*8~fqR"BQ&)FenӠS)rF ,K*99eiFsRIHka -V P}D=5muRò !Ԅ/h Eg}Ԭ?moS~rT0i$~ 4gڤ̈́wc. okɀuf)D4Z|-?xn[' c EI13Vk2Eotrɳ$EH! hqb#8qg5=lȹWx$-T*d2 LyBEAYevT KRs4 |VYiچxcҴw/Ma\joki̢WZ0Ů{\sp8'(ETBsR95B{BmWX9p!<~ <|8^U&.!:] ke* *{Бp%ԭTg+'Y3G(aVB^2H9J l+T:bN<}=BQR).$ XV*4.PQwژE yM[s^@)!O BF#gl-,Yj`kR+F .R+T ZjRsd[s@ E]j;J]^l$=z W]=\͟U]=Z<Fc8\]۶jZq`RO> +W'somgHFtt:Qx:WѸi$k'QMO;QNUaLNvYze5w#~tD 21_;>ZhM|=N. ;~3~eUg}}31ҳEẻrqu4֏EwR:Kke*8'D_.ص}|vY;&vv Vy^!\Ad ** %B 1R  &rs ;.Tk+Â\yTm}Ea}kVɭ|%~TrceKiZUң'wo~Y]Z|G^y%w- QsZ_ `8q]N?|~߫sh߰:G]U %2 -cEw<}s V-87g.Z* hbpVǿBlXbc(7p_A W[[0U XFrWh?o$tvCF0, 3UR(R,נ&*2ì`,} b((z;ӊyi7 &SQ9^ id{҃Jy rХc!,C}&1EJåYrgV%uq799_T+ޠ(zΈڕ+u6Nik=G]#j{ks"O>{Xay&$f{zcx1j.xRs-8(m!3A*n l vQ/G-e٧V5?e(^ZеoՁöEZτnrj;[D%6v5nr.irlC $2m+@/ tlHS @ D`q'Έ f@Y1$ =`uƠfML~҉WRK@h4ror inq,Qb2x mf'ZmDx' ̦YQbӕy`[(7f m1QZ9ٸ}Crjlj0NKV2L@B/wTYGa qFh&C*:'#KY:l%9ɭ&W^rv0$Eu9`QYPVT1YKR2Q1ެ3]#H- 4 8&u`40-"iP2dr9Y-W1^_èS&" RF-9!D91'hy$|h&JЊ@"ev@O?3; '@pZED\ÜtNq: dAi@FS-j jQZ;i&{^:{RYxe nL:s  -ű_≭ ӓbMorLSDh:$L1b-d!/q&VښmWXǭ}Xd[ ߴLZ8xּVaQ)XYJMPRy̒Fu6(30V}8d[ KxNa9XnyGԤT&7spq…;k`eQv=.]R2 HhM)0OJL^B< LULVV9M;I @iG @^ɋ9` U dYe4*&es YE2EgfR; /6^uRYyD.K: uL$P':C2yS܏o>}?8;^-\:,@ZY#7EC $D9DQFe GsN/Tw=$ȵff,h(Q2GBTy-:PYC(I$g.`@a/{< E[ΥouNX7a`ZHړvUN .BXxi.K-?z-?Ѵу.V:`M(3@riԈ mQDB`2ٕ$cLr3pDh`r')ՍigJwշ`09guUjɂmV g쯜38Hu|?u>:V vX>4oD3{dtu=zj0fA푫+(S{.$fuQix>1\w{ipF|"\l+~3u_ Iv'Ҩ᥏ZQ?3^?-!'@dzOWY\5~ e/?}F|mxg5s]28,7[c%;:3_G~8ċv49sq=WMޛ]"9X\NQpCnF֍9-Ѹ4[>a/[v^[nv?_jo=Vʑ@-b0b0֭fY>&cVbK~t7cv/usz娂u:UNQAJ}y"գYeĔ\JG`QZ:+P?,wq*e(EU]zH7wZ:GǗv 7K5) 6VOOw\6}W?(OSnt{݂&БbA;Iw1UF&IuNia%bi oc'v'v&xh-U}P98m4S#!NQ:֦Y!\j ӨEΒ9 NxY>$9@r],7 ".y,2#::7ٻFrWc8". nd$_EH^ɞYop=nIjIi[bǞQSjvSb.(#r,`G@-hJB+j&}Kf:g#޹S[cRR`LNhС$r 9'm2+M}t*-A@G 2̔3r(zx251HAfQ;brE%ZJrr%  uUB},> [BnVN$Y#>lx%1"Հ6oՀ3>o椙7jrx_)Cf^xrlEp>D}FW>ӯfXi+ZNd k 6>+$XD2yiʋ(ߋ>:/r^[3~͢oDtYe Y࡙!b9gAIfNm4X '-p%#$\TGg:zzg@ y^sndOLicI %U>$8 䩂<8.zy=f]qgI1Ad|(e-cF,YH50)h,K vfFQѤo'&虳ġCB=նaaGY/2v׶Oua< \L@*f=Q2ZyP;";-l<,]Y1 9x-#}\UKC6b#b14{tvf^3@6me~?;kB?E$K%g~6n,q*9/177]jť76QMō8[7]4;AzP&X$)ϤAom2'uT|4.~s##n[#n;ݢ[ mt9jX~Z!1.N'Cs͏MQݻStosmm&6lmqTu۝Yh] xѹ%ڰ$~%k%rmF]:y2[FkWk$+œ8q8 ,n< .~4we:iºGpl2?f{ҥ̮.8n{tERpyְg75,tFmɻPy&0o(]ܰѺS f!jXgTbF\| K2&QR<'$ag$Z3١ޚR Th|B)=x(x(gQJJ{v⤸JZnji^`[B(3Y4,S&4:IHhs~#zEcܜYJqR#ZOR|Nae^xXuiUVzoJ4O'_$b?I܂}U_&\*"Q3p7 ylF&a8D !b~FyO;TkUtLvmfPB>&BsB!t9s@9@!r]䐜<ѻxӂ%42W9&`s2O̿* lUYߦCBɫZg PI`ZҺAbj|uQ嫋ֿ[j"6^|C]A%/~wG2?M`]4.w4q{y]viR%ż-I6%Ds̳vg}Z8MBU/0$߂ ]5CM7cCհiGVoD^ {xOE[Zب22n|jE-nۻdށ'?8vf]Ѷ,u;w|qD8V̦_J]?kcn^WżUbd6jVo+Q D`U\lުFI!dLS`rp>Z:z"^71M'[ S ⒤!8T"3Dg3t06S٪&Αi:l,k}['|ی7o\?c]\u0~g1쁮]Oon>p;`ܖ{E3_á0Ab.&e(]it+R;;8?иF0絖a6oypBybx`WtZ3GS=b+\>E;w$S/ l/С-JDͳ;я]̛K &u\(#cN5qWjbʦfU̪]4Y{N{Ftu%YEG毳/|y1LP2]63)Qi-K Ӏ'Tu%Ŭ̙ Mb$]7<&hX!PZ+8hj73bi:xjz㹫{y|SI#}s\2jD|rM& 3`2*ej$Bhf.O`\X,f6Fzε@62V02g)4T,P X9'Vۻ 9]<_=W/҈2ܩ <ٚ \mS =릉 a(GdBhT0 ph/e\R2jW%a gڵTVQ[ =0}Be"-)Ռ88It-Jri) +ٺ*fm,قd.*rEVt`bш$ "؍D0kdd9QGP\~2"D[&d`Wh * Z*AFmYd54?&$x|3VmjpK 6Z"(g*Idɒ\E"3qZFo^u \/YMKNE[̀.x-6:[Z`c"֮;"*\|\<<!TC8 *Fm7/*n|pHǁ|)8-S.o@rT5|'('UvҚUnTd!Ll!M?Ѵ #C-Dk3p` A;.֙`[dDU*&eo5YFL`d$,xΙ 4qDzъ)6P2@bLQ,gM]:~iUm;U5/NYyfHߍF񼮾.Z\ +.JHg/~h Zo + HuqueziɬȕErߏjEYb]*nլߗzԥG{ 4'a{i؅p9}a9<5ҨIHڬW&{`Dg|~bj5As_9FE|Qk6l7+7]2:M㢔S{ܑUvNޑ5RrbqrX9s_V;0(bޜ|E#61ptVӅsڷq:l)0*N5 p~@ > [ԯ)F]NhsrminOS:?T: ZGWԶ5͋%[s͓W/ QRdi>4{+^ōOZ&[o/c]#9қG:Ft|01-tVe XM'竍^\9>/oZ7 vEvZFUR͑,+"d=d(2"ظ*JurW+,Ci"fT~ GE,=eNV՚8gWw+OE* WFMUMg,ȂW=nꈎvjYx<b¿>|owz w_{_/1)xe=_51φ:)pg,(@nqntp8 ~[Uˀ Ƀ2vY"S&^Y&FBr4yN3mŘ> gMز3R>Rj*t#/?~z]9k ;\_,+1i(jg!xe:kǵ[;8h^|}PܷHR0b%!=ġmӫ>8m㈞|2ӳ+kJu~n>>GM_\V/jtԤ ,OE#|h5nsc.kjE-Q蜴TO泳q{z~W [evyOΗ1@1&*!WJPy< Bqṕ >y)9* "B.d4"gɼQ^{`Ys7mbCP%f$Qy 1mgd? z5-*}֠dwԥ}6`J $B;;9QCY}K8zmYqH99~AN!k? /&EC2`8hOHk2$bJ ^{}okzݲ43 Y]УNXsB(^6@Y!^9;ծ~<ƥe3 0ȼDYm%IVQ[2w G !c0 'o:z hk}i4^<+B:*bS2` 6hT*W>8W)t WV(F+Rzݧ2%ԝ ^)V*&{DIYVQĩX*ۀi {o/mҒK'rV1I`'"8<2x=!G@BR);^GnI'ŞMvdu] |Үԝ:9ȥ.wd|?7>W2z8БBCG0GSb"5@j[uNU< I MR4VYJnIۯs*$=I=\ٍIOJJ+|\ VBJG=2W iePր-HAv^sdˇ퀴iYU_A֕lY+Irr̸JaU"T3a_%q[1s/P6jwɇ@zA?@Aar k0o= nU&(('h`f#ADB"C boӹfd|[x|7=یAU8'QUlZ#p1y7>~ ~b4(I =qRuB9qtӂT'F0J%āy{a=v5fA<{Eـof7ķPzQf]ePt$@XS 1'w.FYPb;bQ`Je)&,!xgV1Gp93A8YabY]hq*1 2GPJ]y!@9a:XI}m:C"Cw)(EB8cgiK%Ej c}Tv`r(~Q 9 -wxB#LekH뽐4 1u^k_qn>?l_c$H!c+dI)$N|PA&xu !#wFa/g%phgҤ$FRX(Yg>m:[KKx] ^K N$ü7*&3(k96"Lx"|hK%y4d]uor`H#\cF$DK f0tAzݑ*Jɠ' z#Wrn& ΙHXdIbe(@'i9V BVBYfa= qGhwߣl3+fPAWYa<_,J L1!z8 3CCM@Qߚ8>AjX &^Cnm_[G\>s1AA-4hٛd&˳᥊^Y1|r}kw_V!u@,}h:9l~^^)M9p'xSR;gG&`W<&8{+d[?Hum9 %G~aFy>;}tH_<|ۅ2i3 Nk2xSՕ9h^%L:R\@?rm^ZݺVĖvmQG瞙gIHx_žxi5f ͷV\^lսkvq^b%T;;ڎ;_gW^m%W{Nx6w`53_iiyio-ߚ_/{@I0d: L@X|\8_,ڿ;4mZ׭jvمZuͽS>yd>>k1f}mSJq J^9wlPv^G혷@EC&nQwQTI8aXwŴӯ"\&ч?X6jٸ\m ܿ๨D1\/ݜt MFIV@́1%#I րRB*0A1b,X(;QGMOJ=\gdz?{Ʊ ]외@r6yZ}˒MvE"eQDĒ.bm"T clHGPM芸FoRM0>Ni`'qU!zLѺßvƅG龜e~|kDeY:f?Q&Q;{q'f?_&wH"*`! ufYkk$/?l>jSG%&Bu+DTmֹsR"r1`JJ]rjLg͜蘵M>i.U|l}bl'nU33&>s|L>xU_;|;-GօV-NgχqQR(&Y]1#^}ʡj>-@j|*7ix;umsU-PJZPuG USbYU#l&*Bȭ<o%|]TbEӲUއ0Az*oXledJ4$*'mw t(oSk,:VNM`3AJNmm7`󋺭gk_#{=nsFy 1Kx~5|:0GLIQaSp@:0dES0FW#sM7Ɉt2TJ(mXql5JHbh=Umj~+o3&Ξq_k<6,L ?/;$ߞ3䞭fmK^=.#}t;ٖw~K/W/}b~ͭMΔkB*ڢQnڝJ/Նnπ^%ֶӪz7Iyo/{W'(-ڙC#ۯ:; {{0:h_]|tDžjk+ޗ}o]7?w+;rGHYz9U_cn#7"nn+|Fϛ>M\\蛟7)al~=6?gJ ytgg6ugat& lz`@V8q{a<g Y0:, Q'Kdd*y@S 3V)̤wpѪq}LT2ʦQ|u>(SV'!2b] "ʻX@F% IQX#AȑHqlme^Vx .P]<2ݟrgղWْZkK^ҖhM*D\*IcWR$PJخ1՜kb]AhSR+@9̵bv&hm1TqR=c7qv{~X/B 㣋 o)O8ٟ /Nw_=vLM (`ANcCRbi6Z 1JEڷiaSQnC#{!2160:!3𑯀w5q|^,Q̣ڽP6^{Di|B̪"-b ޵t'\KFs% Umm5,+VHE,Pd E%89FRڬBwzI7qv9_UGnXc.bӵjdF{&SV JnuxK q]F^_ޖUpe'ڼzc#p>}QO&0'vwӺ;I MMJMܴSf•9}xw:j>Y8KYcU, h=cDu;1 'gߥKLƗ`C zq1q@[xh-y^ffur&>~W}-_򀥂P-7mJ|*PF! x`[k3ZkE`z6kmsYk&%qZsM`]5q jzwJ$ţݕ!W zz~Ҧˇja:kj)j,y~gkϿtM^ΣpON[6ːSrChJY Wks6=?=-WU mSfi"(6ہ_yja0džmEhױFyǡq\3f v(E|1WT yr"Eu/}?>Ӎ> (w'rH96{܏s!#12*o,8f;-I9)CcC1qCH''@1 El*姞}TLq=ИcYW(t5Q|2_)z/ Se' AԔ q E^@|LvǟNOGЮǪܰ؎)Ŧ;ƦiFv6Άl4iׂҏtgIF?\w/ӿ>ܰTF`%! C6>pA5 ܶ(=X@ֱҜ|pA~ڶ3_фVX8};jj\Tun( f"V1lPp- Ԁ:$pM 2TAT8 Y˖Ag_U=6g/JYߜUg/xЈT0 EI6T 'Y_Kd79xd%خh:[C#gtlvl؍ݘ41hPu7bolEiYX`P+397ƴMaD>stmNC0]tj =aA .Aɤ b3UХV\agEbҭyp]H1nB"q1X8z<AA;Y/.M=_="&ḵ;Y_U|K&4Ț_kܵ+}Tn<|˗+ntW27_-ڇP# yr QX)+(=/Lij+Y:Eoo3Zn?L۳g% $- ?sJk,%hg (jSuƜ)YT%K 6E.Dܪ"K Fʥwplgv 0U0-*Ö+1kqaLXkF"sBQaCe8@OrLO?{HC K(tu9HK|ꭖ # tG!UʎPJ#OyҥwsQ֔JɮqՈP#n&ʚ؋OY7tb -㘕t78,ճw#}1Xa4&G%˴q Q-ykB u3QFk#"}=n>ΕTPӝownX2l.fZh5WK ڠT X#YVJA7 ;%;4P&Y[XS8ޢAN N Gߟqր9h'1䉂MB?zbN(f[0!\TXz';e @&lW9;v Yw=lv#<]hQ H*')A6  AUI) :W V^(ElơA(y+Yc&k*@M|SϏiYO^o\ϲƣ=fX]]5+M 6ida^Ӆ0j'ˤv؜@5]7ty&i8^6ݓ/  &$DXcmLBF"H,j8 js=)ge:Vi45cG#0H:wހm+ η"gCf4^{]{o#7*Bö@p77IvA&;,V,K$_%K$-[t0gYd*փ9cgnimrFT:FT4j&mUd/B)D]FzXWeC{"uAd\ږhE "bdmٴ1-0٨HA?V27dM7+sl_D1VsC]2j|Zר*7ϫ~NB}c=eu*ːrj#_mQi+KWZ.$feZQqx_PfṔm.tgM"8(O{N%лpe0r =oa+٥L Oq~o-\\0?7_8.N0g_w/ 2jgY]0i{"98lE{2&'ZqNy8qiۓ ϙ8El(@Mz5]$Vy&n&WKqG.z|̌eMJՇc B]W xJY](_-гv6歽Gz?I֦fN˫$vҸjHT̩nݫgK.꟫'ogobr߂+{jm[%[\o8p=+%[~ׂoI#]5 Ú1qpTy3X ލٿ϶M̼qTF:dӨ~6\ݽpheY!!EIy%P8+!Le<=MZS)=Ay] oT-UÂ[?~OCǾ6NVh"3 Y0 \WO󣡢J-SU V{?@& ߾ݷ7/>GǷwo ,㰁+IQ>ooh\WyVh_&Gzش Ws)j8usO|8#6\F#)I&$iU 9 xgixrK\a}F 4F',,&%Z)Dt؜MɩfԸB mkO߇\x?Ig ]z%%wN/O>oɛ;0\1V@r K=։(pΛæ!GNttnrʼn&**R.#,s`΃)@DneʦnW ɂf^@9m^Xd"i% Lhˠ5rvz5 H}npZ Dˤ>'S$$*1IPp<>  +S<.(B"E$lV&}N#Gs(j6m@*# 8$M3o pM*"dVKEpIί ;SR`:V/@!߁[ŒXLknh-u C ydacGcX]vxfnۀVƠL@mV 4φBYjhv2 ) + FF) u00,hܐAh{D.Z#,Ar #$iHZVٍ+)2rm6CGr"nޕ?}N7SW.}"JOWTd4:z&td`hjZ9K"b6IxHH,"G">Y&@c@gӠ$!, Ϟa7;j4v8+&" kp3h $jtUXj]S[R=OXx5>` qqJWqr_io\>̮'i%|pUG'UHt^ ~֘NwB~AjѥY_#^zxx=Qn* i"opc+8Q$k&,Pcx* DKpփp(]| F9ᔕDP< q[#gR.j;xՖV{v l6SK&ӗq*R.>^>r5 P}TT #)isTmzZD*R!I@>;&ouJ/!JhC1prAG QAgiނ<"q0$hK%T;$o|&=Aj8ϓ|@f~^4ښ+t_pGL{Wdspj4,n~aF eBp k)sf̩I?;yS:+Cdf*ENP"<.`h`B$rsDii_ YcU9xwV ۴BzvQ''Wƞ.\wH{Zdp*3ͥxtsze,Bd@D]™hep#fRtTkdP!Sq-spk8d4 ًN{ܢ=6FE4,TWEċ)ś5⥔Ej$^TuxRhEگGwu=ǧկⷫ?شC ]x&8]hFg3t}Hg:zCGNHIEI;z0B45E%R+Ƃh3_1%f9z/(Gj )CN'.˝^iTjrkidk0r gߤ֩$ۢƷYU1[k|w$65LJ;|}>z4}?{׍lMŢl2,0.l"Ee;b"l-J궏I-6SN]-Z#"76ۨކfm/Ѫ} j3I7Z|jvAv7?)5uoWݴ|;Gѻ+#7CUoyu{mog;⹘| W˭sCpWV=KNš/nwV3®秣 6,aY\FW3P#!Mp1; 澋pJd`;ɤ-!xΊFq4ٿVbS:}/ZBuҺD{Dg6v[_SP25p*xV&{o} d%w\__(l1WANq3[ЂB4JXb JʔeŲ]rI$Tle#nS!T08009ũv#gYJr vc/$mmȣjAXV/griGs&!J%YlELUi,URvEFbWs1Vs5zJju\>B%s\ɤ-* *ݰg싅3ˆ{ik*/耞1M7>`gO_;&M (ANc85UDZ im.*Q*V7MEنf lr0೫Gj*U%ǽq{2ڽc_Ψ#j&OYUS@Qג\*ʊmC5,+u]W<؂b.XȢj+:krQ'ű4Q\mVIBnlÙQlрP싈3"GDX%RbL@OE,؀TA\HCQ Y+E6.jh_IDAi8sLNb $44Vw^lGEnu\.#ٍKEi7rTK)pLܸ݋(:c1x0x؋;C>u >_T 1yrxDB'w4NB,B6xB%hDjz8"U#יcFׇWBSjdxF9QpjT#\;\^*Ng#>wU# l׍pyԮT~R<*p#\}I۪Z+]c_'w VfN7O[tMo|q'ʾ2'jr=r7.,c '`_+xl^VF{w}v1[4\8)eW'.Kar'AṅeٲA sD gdz}n{Z ={H}[*~n'mUmqvʘoNgo_,| 1BgX۝ڿ>.]qwț;k"U r CV<'@t#ЬF8z)Í\En>ʰPP•G ʀW+!X:j},pըWJ#\}p%.&omYaEDp5ml)ghN]ziфӕ`˕z?˟Y ?goC )sBE;SV΁ kjrqQ^qs֨ͱ,V=tܭן4'],Cge/NV"-(OZA700cR.\C*v8X1l)잣m\L&oo6aJ_޽ׯ>K\rhm߾y0o*w}&{͋uh=*LO<{HYk0D"٤}2t}9N;{q@eyO/˥ku'ܛPY}ӂ1X;X浕Zg*vmF͠i vx}?k_S.ݫAϻo~t{dMBrh{.ؔuD(.&z9̊tb5+MxԼ6kr]L;I~ڸnJ|_8-+i3-[8C>pZ.Oe\{dA\f/ON[6y}ޏrrMJaI\qqަf~ivή]r^u[(9Ϥ;yYlF2>77,8g5O1pS~!|f~gzIAwU/_N_nB6i8E̙z\yqE*MfsURi"=v?<b& VRTE!\)!hs(F+[X1%jmty_QSJV>:=3 1:GN ]Z"g{mAB =Ej(6ԕ'/w= įBǨb>i^udjҩ1惃\ דWѢo_p#|4D74x@;!QU?mX T GjR:Ԣ( wH S{Z1[ѳ>@}O> LS{g+j4\B5yBnulB l()*YFΎbH۫|!ʧܯ^_gMoMyOcV>v{<]qGlh"A@ GU=d7m6edJ4Ncؕ]%8ymkXa救Ϸ}]x~&ws1[#஬ArˡsiepwS秣Yگ,ѕLD xd|S&L02 z`hGk5.6E1NX0(d Q'Kd7Mt:oh5EOU!c[vSQ;Pj4{k ,yϽcgg)ʦQ|%'9 -(Dt(LlhFqWA-)j >[k+TH:"$5̨/rtNq,.M{9_}w:NdNsQ=v^;/oxVL.hB4Dv6M* PJخHj5F*sF@OIغ\}JZ1I Z[ U@72v#g;2Uaa7 Mg,  <͹~4W4z4-!\\L>\MtĎiSFXc1!pMU) ^h]T(F)T4}o>.  44:G|v54W@_SMw>n/r#v\`PwqDQ 1jhSҝ4{#Z2PEYʣe"x%늇Ւ[PE Y4CbE'S`M.*D8f7"*Q]\^ҍx83'Y. Ǿh;#qDč["*vYD LekĨMPD>U@ߊ-RPl"DAN31X$pJbI Ncj>;#b7r#": .NK%댋4∋9Yb %rfjKEf&ZnzbbG\| \<<!wC:wc/*qSLя mG6<ڇ8*&U˃S1CT VYxABsp{p@{)=3~?Z_ւ.P`I1AVb+ХZagbv*rQ%vQH` FV-+Zh1b gh:vv#g@'Ѓ9AJd$rVrг4ș[V+om5@]u6 IBnJhOśv1N)#EQX)+(=rL%PF($w,M< -QTj 3ĭZ$fe3) TؘKVu 2F.ܞ?2+"@,gQfr f&U.?,A1k!`H' k9d¦ NV]q" NC K("7:GJ1up(5Sj9 1bA(FMJ4I7{id]*-'jDPB j{BڱiO97XA,8Gwoa7뱡`f5]1%t$b V_YSY)7J[?ȯ(uc}Owe.qLI6YPnZhtw-mI 69x~?w>Yq ܮ S",_u)DQ#'1li*g 8p 6Sn5<"aآx^fnCvҲto z5I3`hR&x۟ޓ?2 HM^(D~@RVVm)%FtǒIK-.50e5*Ф(ӳjޚWsYpf}7oS~g %' /x=@c(r06!Jk9E1EMBҲ*\܄ A2 ǨU1k |s zqc[#g`s>'W3f?N{ eGsþ xfT)AB8lIK0UAB 帒H/:\(jY)Y% ;"E RrHd 3âҌ;$ (U)a`ucf(`0 Uїd: 6o!Tp>IЮ]S T3 V{WjR: ]ϵ5mAR%dpDi,K< L婍l$'K\s˻͚#O=xޣ|p"M%i\.0Ob ƝFI%L. BOTzB c\{&FlT9,~˧ ^y(jլ4w:O$$cXv.bd.`2} tYߛ F-Ni&\:$ULlAu뻓sN4 =]/1ɺ:ȯe=_N֖.-}>:j'.=mZy<Q{ D[k!Uݔ&h)4SLA(J[ lCz Y )Oc.*5XMn-{Z#~d>q| eI>呖cN7SW.Yd/bkEO-KͱCG/@WaGazڗDDLOo |N'2^߼|&o^}+Uåh|(ON0GЋ) ]EiexD.}\L#5(K68PY~B'gC+lu BfF>fɖ#$uU<ܶ݁]-|Y^N9*`+ȺvyK79Ω:?|f^nQUwgWR0@q᧴RzUԭFkNyAX܌sR/ǣt_z$Of9Ž;ܞk OIc 2gH)ARKJkOfxg=ƦSCƒmV)XZȹҞAKœoc+MY(; X t*°BA#IbiXumӫ_!45ѾȣNMb6MnZ3ۮM7!zr!˹}PKEc-m+ߝ[ig ͐ 4L߽&mDR+R[2E%AZᨯGmdS_.j^Pֽ(n@r / t-oz!%9ɐKmZ~ml]*$TɂĔ#"t[] " ƚp¤B!6))+ʂȨa)'RHu?d =!"}2L"@Ԥ¾XYh69"N+ϛ1B8UaFs\}k?0!`J0 \И @+H$`-Ah=e` gv~NRi˅׾]W,N|p5hߴev#{!^޳373b-8s9jC4Wm It"6x-mUa{tıx}-5O5(8I=Yxc z>~9`O+z '}fR mNu'3,o-Wzo[-ụC7rCTyRo 15W"罾H9'I؅jv>a%~wY0t{RҎ(`R}$EKj $UĊ+T>9W璼#gz IN _Myn4߱00Ϟځ,.wPT;\w͆{\mt31tu'& g8nu?{ϳq ocRTM$wss@3l=`Ӯ|]h?w/~6`dq `6dwRe,!Q<%[Ώɞ}/shsqYE/b3IĖ+iLP&LhLގKip~y_[+|)Ӌ˜ A>p{"q몽58?ܘbPJ/6x|/B`*~ʂSOJ3u (RdKJo+LwY/vⷪ0ߎ9*vw̞(O 1QrJg{9W!_|O7S'BW]KFRMl 4vx%<^ߦzu= n7vzbRJ)|qe:.\Qy}ۭO!|]>j4ca>Æd}>1n5# Mg~e'Ϝ1fp}ǖ4>^T ͬ_.檄@mx6M7hUZRF~knT5h6[tj~^D(@90GcM;'M[3bVzc2`z}u@/M,>ܟ/R}=T 濻Hΐ+u)SrւU~l~}z+` j>=vg 9PUs;-zW|@@gtX~_3DF;Ͷ=F8s7UA|빎i|vsu6Q,9$z^,]bVy&-'WK| 7Ӻ]OcT5ꊭw1q]:0uu}EؾГGv~Uc~[9tj7韷_.MVGhYgnN;ˋ^R'^8ſKA@6%vڂ}X_\t;5)tq-2j$+GB|UÈÌ,fY>`Ydc^^]U+GQ'\5d\ W͐.+*d:jURԥ %)tC`J{ RS1?"}=Tvշ8t^p9j\>)Hj.eWhŽ-FTiTI/)^Fw H ߕ?w?qa>駏߿'8zJm$X[׭0| O?z?r^K5H%NhNx-3:Eb2L29,ʡl34LMC[P]UF2sR.]ꔛ\=XZuv.[2r"^ʍPcx޼Ee 2a N=V1 ZN{$ >4:"ldmn[^Πw gv܌鶹c%VÛn^c^}1/1ӐTP GU [5t/99)>~b_i[/?|Z}EU,a*C06̫4zN7*i|}Fm|i=se1"' %4p6l`*LC,J1لVdވ$cA `,uiy,XTH_I#k9/`TAf3}]l6`z9u!ėn\v5.lϲc T)n '[ H2ߠ~sVi Tha-*코ӐغIsi"͘wc}|8swz!]W;R1qGik=G]tzgKs[Xh_.U~f9`Z$b\.l̃6\0(5^Es r<k1bl+>vWx~`ܭYg#}6n}y.uw2'~}z,vrƥmXA5ِK̑9>LKHQRc<)%# 2/lC ABcAEu1H `ea $X @H\[PLpre(wjGs֗#gQ\ё:G-4AMخNzg0:[?Mnj`C ?u5}-وad0rߣrوy Lnj1\Ҙ\A?d sm2&-Ztj=l]Bl<:Dz맜^m\y1ˋI,1tCzo]®Cؕmf]Wܺ[_M}-턣}Svdmm۶ɽWM6F,v>RrHswҦֻ?b?<)5 $cDɞi:4ց,.k(jagaɆ9ߓz|UV{Mہ{&+w>;Xvf'l DŽ턷1դaSd饽]sCa=Ńy|](ەYlxО?xA}澐Qyg[M IXZ,13y:9O> ?zs "JSD`s[K6ι,DS'%PpD\;e@/sOgՁ5J.;5JHFɻdz )L6wo[dKQpvz<ʞԎ{<^U7a͸ό_1)bLe7/4c|7B|2;+o+uq%$ TX[@}eLA@^c-LTEKHZxücN22GpI[YĿGӞ3ˉ /Y< ,3x*Q:ea]'HFbP<#gcPzXz^!'q๐{߮{GԞ Ɠp^_]iAgr耒LRǿ0!mow/ lrhM^o͛c+̋pt>#WX{푼{M^{ӏ7OFqD_75o>H'DB,\/2R -"Ig$s 9^FN:El@\`% @TAkh"tJ$74OnΝ:whghٚ⽾4szunm'|=I˫za8iBQFHՇz\~*E= Nz\j@ ZzRFJA*S Z"瑹\gz(.[/ 4%O¤ʝAڛ?:&yhdE]W1I,b^YüɽσWaa 7b4T(*ʜ2Jq@kʬ2٫y꧴,TNP"Ѩx2?`$PτIz}g}t|!gY[Fi6;V|W>Tf=J=]) 16 &(Gak8P8\ Yi? ʢ ɁFf@DmB4[&H] EKfLz- Ž-,,+ h:,ŵz\c=vEƮ-g,B[7xQ~D{i+ƣTX]iT[Ho9Fq~V![$!3%#gt *FZ0`MD;u6.7 ʑ뜡BDPQF'S@BntJ; rl%X.Zlq6o7n Mf}nTe7̙0u{ o?zZ·Kr="גU6+BsVHߧBytՌ.X]|םU]kO 'PTP 0b % eKû2O}D3.9iMblXwhE}7!4Y)'d[r^/9"MEQysp RsI)%&dLhvKJٌ6 Q"&ɢȨ@s$WP!qz%uHƱ`c1(T J#c1r6#c9]]BcA_MofQnķثP/=Ԉp`w슇Pa7<F9/**{]\Lښ+8k [?MQi~߅2ZIj5Nbquۡ w=}n;6ΘwY6u]q6q& a \beS巨?hRxCkƉlv3=;I~8ap/sSapϱ1GkUc`y0g&;nQ*0, ?TL.Uǒ S3B06pF׃o~ۏi˜]9mAY;]ϖӚjy$(0 u@ ҁ2eRҠBa Z+HTBq#8}̲b!zIuڎ'YborwmYYvNNlik'ŧ#wBe.\@fO[!4明<$X92ˑ77t9{~=Gl@Ur^$>go%[I5A1w!,CsD!ͫl2VES!,N1C3 }Hj4J5}R1ssҌ&CO.\Wy qάQ6;BV 3i0INC,A% Ы0TBYrҒik CmZ[o}3@V)ov3|A͝R 3djfN~}5VKC;r`V퀢>~]qYQ[+ 5eho7C?iQ?slldC>9kאBspIJm1&#rΜth@BZ3Is^LQ죒NU\[=1V".J\Yڃ4qR^&Pd)4ɇ/ n PT-,pyvoJ o|Wq/ŷ~tߣvho\#JPB&DR'?%mmO xZؚQjs k"H<\I:I)*sY=DsoPyK_U+a%`ه6]I|* _* F %>Wx$cJ Lvi/E) }(y Q_2۳",fht h91S)R‰u|>[a2轢`C m<0RPlӠKn)E!"d)}6 *+*-&)7\!.WyP'+C&=T qLRV6,: ?yoר4%;hWivV錄5t_,"Mӫ(mϷ.UhL`zm`PJ[9JSZi[]yA݋%||խ:LzdkފGڮKb[X|K5k]'Bp3wIVҽ>J-^r/ ™y"jr?OvvV}vL0ܰVKԼ)J^=(*+.I׳=xJڻ,Pz^\s&(ϵ4'TeΜA5 ʀA00ؑa-pTVq)r\=Vv\H. &B] N6ѕ,tͯ n}: nlƵYTI;ԤE֕'߫2KrS,j7N` *#yƾ$T25 <#K/Af8s)RfjV=\Gfxb*A^e*EZ(@H>H'%D@KR`d?+`'|:Km]X[B랄8 NK18MG+fq2.U[gQO¤@Z5Ptr*c2`npxg4љy1ÖXcCoiBuhciښhVUm5%yƅ;ܫ ޥ*6rSze8l#R9Cޖc֊`rիɼTzu^=/H}V,e2O] K 2x%sfeS~6 k6by+Z4]{JȢpߖu[`<mާ| {6ȰDI~K^%P%m%'2{Qě*R"wK} |JM yLzMmKh%Vt!9?U] 0>_у֤e߶S)畹wA2>Ds3\]ft* >Y䷰!PW06H }i }_Ɍp"7̼4y~5Bqt8~n@p-f}aR)}n HIR =3,YսJݼvx!J8y6Sae(=A{1_^tf](`tNfśIU;ލ֋{kp,JSnØC ǕR]&IWOik&e]첦GbڡبPrÍtFדЭ>5 Kn4`dե6hCePwc޺$LTQVH˚fue[Nԉ>Χﻞ37`=[TFMQqࡳ1*}dF)9K}O֪t[zQ%uUdsgrx@+@ @kx3M::P/L~!ZGֱ(LV[)L,= >?=ns[mK&%w:g.FLW~/;v3"W>}ڣp[.'aQ{$O0d_l{+z] )z+z)zl}f5+ew0ռݍ4ܦW<;EN}SVR& 1A}0(vTQˀ_DʜQrDILj;GyO~rCa(3*h2:ڀmT 5T vL[w6sJx2Ocxl=sXSY>fIPշ?IуȜCqXΤRFztFqW5ǧ9ZI3"HTzAHbM)6()z^q h-#IqƝvm|A{kSBR!)J"ElI=F"[fwMwUWUZHR!2Dyh"(Jg0,FΆ~W`5&r,8Fٸ;Z-0ڊ N7oᚊdmDشښ>3@nk›w=<4}`售v{Ik7L]}2Z_&WG2ڲ՛j˭uJ}[9wjĭn;=_7.QrB!=]_:Q:GKc=ik:.ޢi[x)4gMɮ./i4GpPie!eâOdK>mbk2~ATU][ʄAB-ֆ %"Z;[51ٷɮGQ@)x&'me`TG+m *'(. Q픗3K+T>Y[+)vmT$2x9a&e2ZAHF:@d@*zˉ)* XB H6L%QYG!ьKn/ךPXP^PdԔ BX ~}ZO ݆fi]XKJ)=%8ϕ ތ6 QF%B$lɢQbLޓ|^ UId=II8 QQRJf,FzXӅ8c[] BՅ7Ns&>=iDoԮ^{5rKb1P+ s&DJ+(R{"(36%^I4dcEF6A e62> ª M nYpL̽ڥc[- kmiAH4мΪ|NMҔ%+$,HCIY.TmA4ZȈ @+:pШk"0Gv#d$jЮtv1rƨUʊq_4b1V#Qvӈ@*&[  ֠(&t"d$ ӠV(D%-)(-6&%j(Y2&`B@K [SryK^#^.G8E.퓯lua:{9^2%UBc(?Vf3Q'@dDJwz!Rܱ>4N> +EUauяOPo&,nH!ISkD-'j) 3QauÅ<{E𽻱+Wt͸ݎVnv*0'>H9kr|>&":fD3GND{KFk!hqR 29`%2KRx% uεV9g DK2a&jS6UWsRQXufx!$mB-f\Qmb*<wvQI+}2ңZ˸ 1Z Pox'$]N%x8C09gQPxi8p<#6D!).}:HB˦wrEW3aK(22P#r;| hm W!0тdP;(8 Y1r6F Ĭa}C6 s(fXPZ%|)E5Evz"2@44 >9]!w2NJ}OzZ) . p`5\\֊.sJtHQd%Ž rhTAbP+LIskdNQɡH)oH'DTH(<4ZnSE&rlX':6DEJ Gl`1Anhq#F{sGl+|knlT HHr}J.Hɳ@ArdKH,uÂ1c{ul+to z1kp8ؤn̳d|MH8Y@BlD&P.&#@Faxv6x&rIм7b2ylo}ʋ7;+&߬+-i$u"M]G~AFj km^3n,V'\bLp0ӃYFÏwGC2 ;]fav[/MofԷn:q{-ҭű:|qAf:nFh&py~ƕŬcif5??kr'oEBÏ~DGWlWr~44v"PvpM)JցT RᵑV#b=`ʩ8][Q.{bW]>0͝&ߙ$Fs=b<Ds$c1Q0k IrEI&9ʺ@"-XaǵKäPHB ;\'4}ԖhI,}Ϭ9>3o6oxrbYn.pVpcC&ff:INXL^t:2A./B|:/9-:+6Z)mA'mp"+dFSQ")59Yp-[+{AcPȸ'0:!r>ۡa݄q+s8C*-gQ_yQhFpKs]Ρca tX[FAlkTE"u(0&RC!8-M<ʋ59,guTDɯ*' "z !DGK)Q Dx'Ҭϲ3U]O:'./-۬NPXs]Whrvރ63 q3 ~p | DTbtUM_m[&9>6q%s_k͚5=cG{G4{6mtB'0i4=.|ȍJ9‰Tj̏Nx[Õ+T?%G)w9~ۻ T#;پ`rٻs"|8ξBq@ [#998ъsÁgoFp7iH;ȩDXNx S9΂ IT߆.B\3FW7-e{>mCeIY{VS~}5sX ܎Os5Zm;M[Wm|6RsMtEs>;9m]UNjIY^"'炐U#tjzcW,Ph# 63g15侗kϳ^rըw֨sux,6eẎ\J(.}+!Q |%0Wqf=MHCUh2B[V}7.ܳУ7´[CKBI}ޠ}\) P ~݋˛Sf_wo^  '8XAzg~މCw }㊡W_s÷nnqK;}PIo@BPD pqFCD6.Hfdήk@;Oã{f `Ýޏtc8 &GAؤDPc 4&HN5u:EtO)Jls(zwqO)7}إk㿞vxBL\1T A]m5RτuD{f*.!]*U@DĉU2G)hE <[YmE@NPyi˵)I,C$ 8$S N3+F W{#@qJA02rX&9 5 #7xDC?2HMvJ/"T..$eg%49uAElLFi)Q+yϑYrֈ1-ȝgKmWD Nj5ɅƤ&bkɹdDd4nG{L*}e,*|g0cht Z IR{a8v2G2#==[Kw-A[s/rgP|"w \٢Tji|sgC-$r$j댩 +kF攭&pߣåm@ףiCr(q@;w~e3-2lȐAh^v`R*Dr$0BWw#%u4qZ0W#1 B2*ZFy {YVH{)CGrf~҈1˩eM Y*Vl%Ԝ8uਥӅUk@Yڈ֋rtbk~ !-+{CZ wŗO*UN'!$X~Ւ5x^ֳ'H;Cq3,OgƯ%/&I4m)m o77gef1V9nƋ9g|綂ԿdSLۻN 뽹;8ueD kfaucq%y%HSGT"Mu*9LܻpA>Yac2M}H;L`ᑪOLQJY&u;2pڵvx3;\qLL9/IFUwN_TG$c'y_^uW&^[}3YZ_XriTSWCw1"l-HC3y*`@:LjTDkX?&oNScn\j(_ l4;sLiK4*J fwN{lQCG.Lijmgh8ŵ[zUe2Tޤ ,TkhvM='g1frTk]܇]<{K:v=ܱLCPdܸxn#z(k9q !C֚gr[/D5}0{uB&*(I/zQ7DCO{SB Őޕ:堭1N(xU0LNת4U̎TFbCԥ+O%t@]]`dt# ^}m`f;.KmOmb A=i1d7|ćkTM廃LeDZ9Q8n(֒D)c1k-k)k##FSE!\ Dd9 Ө>B Їe:jLj4餢ιT!F`<Fy-XЯ8rOmWpXn&L4:@=cE1_ bzꄚL6]eNX-}6K-fc?a1;?ByVRL{݌=zV맫^,D+) YDp*ly`[>RdB$ 0mMRZe r!9[@fZJY3-|3]uuןÕ[$^b,=nbZs>L٪ GÅe-YFPeevdt5P"Z~+b+]FT(HSpXT.%(kt2XN 6g=^?~ ]sixj^1lF;泌৫0)%i p.Xa1*zZ=PcdDXMQI*C1Zi;N.3o+R}Q318u=6#d2fȳevгUxک[߻6;qZ?t4;ՆVz-oL<Lgާn?d'o|b3C#ǻ?~S5o#[[ZPqE{ņ EOݼO6d]|a/ 7an7!nRyxͷ22aOՙS\ Dgf=W!<ܲetuڥtT/ƿb|_Cꀣ0JQPFh C 쀣^UEAzɗx]Pֶ5V}_lBH0֢!nl n 'TQ,*^l=lE?gUW?BR^y@6jrSKAÒvƔb$U*3ReVr"h B nTrBTHsػMgsl1?+^ut| s[C;p]. BW b`v@%!ktP޲f$2&Y4skݻnv5(mO'0Nh#Y@ f  g`󬷹)o.RRV d"C ִeQ"]@9Q@9t (&S+Xn `0YFs٪J,-TkLW('m8kTFZ13^J`W4E Ƣ"Ns+MglCͬN8n/絬&&z)΍jAsi`T3jN={KdHZB-\KO&iGڡơNi~@,v*U:lNeXT,E9jT[?DpZ/Ѥ IJa`bA|]ڪ׈k1 齡L*Cl*Jjѩжj:tAzcz v`*pHnK&WTI=@nP+]hV~cQn"ʧfʷz$gDYPjl-no%uU4234`c#A`J]nFV)L\,_zrqr|cGH)3=!a=-ĝX`JC:Rg=`:d;*] Gj>^DZu1n==K:Iu.$Y`< SkIF&Pe2Lh*1$0p8sxaf"0 ; Y=^+v>_0i-i(f+ނEdd Q2)IVC/:^ofC߈A"S%#UJTq-OpV 5"":J&(.oe/>{}V? eۈӖ2)BNPh] a(DY:Q3 34grm AFCZ:$VK=d KZАmmsK}/E8ԉ" B%A8GRypИLN 'n HLj##hTo+ ?FUIU8-\/bo^Yo?f._@8 Q?ҸueLUBr6Ţ"59?]pʱ4r4#ј+0dxG3Gq`z槖$ZʛqwHtByt\Ϯ|nyZrβᬔȆ [{p-mu٢|l?6%d,[ƴZ>4|h+\/pF^xXruZSݺ>~,λE}Ͳ83siӠ~Q-3]YiYޠ'w?rB(IJ.v2ڴY!G23g1;Kນf?48#[Rj36XB3&_F[MȜd))Ead4!̤FYME6Kpz|m|1UNܕ2vZ|a?洷>Cȵ`7 ދ `+~s x|&ƽ}ЉaA2 l@31GWB[b@1LD-I:gh(у`HãG/`<\X6N 1Zt^{il`rI^HD%8}J Nt++F2w9yYk+kk"IY a]J.AS(F|/c` ʛaГ:\F /ܣPWcZ#BiA'sQx:y91kM`[q@cr^kًױ w!'iB`iLc( C!!J Y@)VFHÆ# ==sNaCՌ*3Aʘ5xBOhWC|(Z8'Я9"*PPi* sEp)(LiB@?)/0M I*NVٹݳZ5HHMoޕH=)nJl]h9)z=w5*zǮ4 _.@r=węY>'(?e,y+CqсK ?qK\YG{$N.KKO{Kצi>څcZ:=;!lRكyo;)#vgXFAg-{+* zwS@81:2#%\ EIQ1tf{EȲo;% q1 e'Ģ5d2T ڽ-S*JOL'jxwc&Ce `[NK%|)zb;Vn}oU`t^M`à[3ix~oCP3B!#* 0;d!J 6<"LO9XoDƂ@*FdB|TG3ҰPE0˓uYϗ&h8~b^ (md1G !ːACQ&A$Ĕ2HbBѝ;&{t-Ztv[ !;C1u'] 6&;0YF OGS}*iƆЖ5zBT<@ZHp/B&W$JDĔ!*Gff|2TqXt><9&#Ғ$ FM8aX#r6O亄/G)=&*1lO_%(35N%tF U/x[4 C̏ dC&.r"mJ!Jb\8OIR""q^젅qB?[`zhm$ZݶMQKȭ_b^ S!!`; ftͮݫmW \R)B~)ۭm'_CNe^nL` řX/#0 [ə`KSW(*L B֖bzA V{ kgͮK&/wsGf  ":0-lbzUzjPjc#峏u*iR`l/dz?P|^BX븜%n&])0-aA{R /:ݓKi3.$s9jbWmdеϙ6-0+aW91/\ B%X-]1}Nj5~XBT yb\T1?Mq6ԃ9qgJ\?+ѳRu_}5]m^B2:D~QNw*^my9)+eVs72>I9ܒo~ڦR(TkHW~BIz*8R CMT/$k@L SW<:^<BJ>;[SkX xt^ ,mzWhTeGu@wͽc>oy60?iY~ ^Pխn90-a=fC5DO.rt`7j5;Npmf!kfdž 慧jtHGIhPR lN4!q2RQ.8i!qt>QC \/{ ȁG!| ,NykU95(pBĒ2i4 -aFβdn^@{{V #ة_ k|½U{ :be1/DxOwmHUv wӏW "/G1,̃3Ӈ|w5cW#?ݏ[rAʸ2nhL KxBn22*s\Ds%$%9+dXc]-&ޛ+҃z ʣWdm*tF@,9*2%KEU[2e+X+6`e ۚ`Y ⍯ѕ\m8d9D$32|)J⭫}PN/ct ʚQôU k) 6H*!Çύy[JMa?(6Oj]6kV0ZrMQ@@Cr\w=OA|w牧\)oy.ξmd>>?~L뤽e "HB3/Eڂ`w\_YڰӘN( Cҷ$9uÅǢGo JlƯ0d0<qƊy``R_z\;)Cutd$g@pmf ft'E~W'@M6p~cҵrPeܪm(T2I kmHBelؔ~;=g힧1z4VmD)[- y)ʊ8q*Of QR*S)mv>+-6yh w4k]ԻfVH) %e76iRLjDhi{c!Tg̸])'И16:{PtFbhOI8_>-&Q_>2ku):f%Uz;j;m1"zԺXt):Q} ` ]&|J'{X@榤gE¥t:RȮPB@&тw_KͲnG0L3RQ/3ޅ*]`5.0`Ϩ-OXݫ N6h N@k!O4v/uh;n2Qh?j(QLnB!VˮPH1Y cMmV::D$l@ Ck -cw+3\iu x0E3Zհ֞F*ZD@ %X_;wMBA[Y&J)1N-` t[ ҀRS̨l`W&4|s'8I#`$`iU'WkUdȮD@,AmTzCZq2F61"g#n0!wD `L `iƗB`@YLu>K@ZAA[:*]bvQ*ACmj(hS1 E@(ͨ#kyGD)Z6NjA!NMFCꕶCw2J".5=() E\}Ju yW@Bb|M6y-Qd)--VcMZPʰ*4Q]AjioZI །wA9Q̈KQE;fD' GcEIJ4B6}ofCu}V.I˅H.{KUеYV2]ABZ$\| xḰKy@xc#TYd :!JrAhRW! 0阊q(w͗,Z茸pΠ=h5'HiD.h2QӪAUYk2!``5~OC 1|Y%H VxQ ox6JrA+h)ZD"hFy¶e@TDn*B%NZՖ7.N&!z*`YedbM=A%DB>hwb>MzC*KthM )C0 zv5t2-`JQ%KcEXA%DEWHPl5CkHhGՌ6X\[vDgcp9Z")YC Ǯn6 3*I@d(nfWbd@BUq@ơ#2ΪU%oCQeXYvnjpM"β]4 D`ZbM[JC:b8i$ M^s2 Úg)DzIhΓDdN/AQBo1&POF~jH!*3bO*Vz7$%xKT]ж+9]nDdh8h5)QJj *TAL:"Q%fTi SZJ;Hq XoܪGͰ%6vł*'XQW6b]MA53`QG|AW\V1^8 R XQԌbwH yT,+~x$mDV\GRPgМۦwZdnfR'=֨U 6)jP%H/e2 b2D ]Ρ]뜼=NKqA^ Q-^j-z5FC-:mP Fx7XA Vp*-]Z4hԋ+Bbx~)A(ӖN;G~claކ`ݨ1jEr7 " J.YxЬ*#vSZ,=vQՒ-* _l95LԨrJmv۷[et;jJlSZ&ɓiSDPY׆an6-!P精ʏYXA'1.B(@@;^q1mޤbpV[8 ˀ X/~j傶oǰ]&4|1~s?,oBِ : 7 Y:m6=UЗmےj)uZ]q6~˿'tW[ 8E(3doXǰoegZLk{Z~x98:7V@-*ǜs%exF]d$ .HvA ]d$ .HvA ]d$ .HvA ]d$ .HvA ]d$ .HvA ]d$ .ȗ낌 O.Hva\޸ u# R:vAТVb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vX'BNGN ?N k F"@/ 0@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; NR` ON \ >{'PFN2놝@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N9uΩ^}39x;t=~58X8_λl&d\2NI?%UqoK@k7.l\=/=\A1OLWkg =]= eϋ2]=ӃM:m)oW( uL6Y #f' |g`L=XgM~|( V߱S)RM8 Fh~iGt/'É!'iV!ϩ!ϋY=D,;a' &}ui׍J=Ivk_|?=u-mjiM1Tyi%7[B^fb|'_+QU۲J?hwHh_,9Ojjj譙Hf4{Fc\'EcƐt./Pc({DW"V ]ڧZza(3LW ]iu{DWlL"mϝ t銞6}Lϛ"J ]nҕuV˸GtFϽ++5ҕsc {CWW}+B=]%ҕ*xtEܻ" ]yϝbztz]`n ]ʭLW/b~2x:( RO\,gC.O/޵5!X\D /N``+d$fy+#r5ihNvkSC6nK^;Ä]]q6~˿'7fѽ8E(3+Ķff8 ii1-VMnh<Ɖ|d };kN)nJˬx .imp[4Dq4DHBt8}:_V^WvpvdAAOTrex[F[{'^}!< 'un8Lz8b[yv-޾_;j]K2mh:e8׸/`j7}j@zRPTI5wKҩkQ~"RZjvZpAg0yA} yŃƧѴ{M;_zgnδhkh7?'gs`2S?X:|}keKhRhEOe"NQ}ɸđniͿ^Zc$1ٗM¯R6'rޫm&iO/}`+5WZTkS=>ww˝fNB-lb׾{:d7פq~m GN',WGl.oJzWvS|3ƀmTN㣣u֢[j]haZ,hm-ʔD=5b*B(UR/tlláw^ٕ}B7l2KK.~z1An^X}wʯKnÝ{G/=-uy:)J F~i><+#~el+<swASex!,.iFaN$ƣ:Emz^m)֮kuZkQ{X4wzzy1>NC8A?EjRx#-*`r4RsOpcCgmU n^~ۻ%m1G->X%W.|&JRoWߒUP=vJtWaps6:ÁϖvEcLd%onhÂr>F<mky!౯euFm^kB_Jo-蹴pj󲊫MkRǗ먦%ZR;sZݛ^?ki7w`{\<_\V۵]A׳Z![-5EcORm骩Pc76pTZ,BOo_g&7+g{]N\rW)ݼxX$ˊYu,JC;qVR^?x=%3Ve4s]OI^KpFVQyT֏ o47YKv[YujS ).Zܫq\4P5GeԖjΧb|IL޾xÛz }Y4G᧣`8~ڞgqZxwmv{VyU^Zi]*W"6j$ x) sT2V 0sZ9ϙz|$#w_Ck4qv0H?My^ʙ؎9ՙs`a?EBҾ^WzwCA|'-CD%g%BvZ?"Q1 p{TUʪ %}媋Q_1&&겘?m;qc奴qUBPIt|gLOJs 垘dzmh <@TeQB*Ky?ɃA{.hn]v;wwd nW>o)5dJK3*KKЅ:ᮌNAc&]J;gp8=(S vy,iC*bqk!y'D$eHD=J  XIUbi:_1:/ ^L(V3-wrSqRb[ J+筭d +V>^6}pr=nV֎.}>:Ԛ ! M~5`bVR̬,ӶE dl.eDqť J[6:_金j!,F1V,tLI,A*~4dF#V $(x|-"6H&NF& ='$jw<}ֆJ8|SwՖx4RaDY,<+ 1m2[͢O.R(6HcVwal~%C:~G";a\DwPo|۹c^@T7R)lelR$@m2lb]J'RPJޥThHx].GCΚd֞lL&ϭpC,{5qL)cҾM- . 1ǹ(v/ j#@ҟ۾&prE::/H];SjUـmRJ9p)qF&LlU u G1u N)WEǹq2+ Ig3h]uTV*+ i+T d'%Ĭ )Xpaǀ9krq8&t`,I:耂y'Zֆ#^bqq\2-JO%{wJ2jny9-C9'i @xn|d"C%g«{:8>#!w/y<$ѣ3Z+c4BrN) ] Jv2VkP+IϘz`"FIXdA9et:b$!M+9YnӪM|%OjZX{4)6@#i+2Y%mJRUČw))]|`ީgs id\gNsrlI}fepv2 >V跛Z)}x~~kzm? /SM7'.V_Rj&w/աd;>q)C[+Bș Ep$*p.Cva) -Zt'%4/K1O!Z9IrG;g6GB"F.ra}R$f @ix+3ʖ~;DЍ.Js[ FۛY{o=</V(,+3A$mT0QI23 –}߰Mf=<_a΂C! @ʝmO0L*BhD]rI`N"y \qP^jN=/,9xW%)pڣHB`R'^+˒JhqXʚo'XfֻТgwm^mˮ XA⅁/m2ŏ8UJr4יGdF-)Rքt92Y,hb$ܦR%!c6 :+@qtĸn@V C3)yfTf&VΓ̠Rg_4nfoK\4b4[h_4zxs%u^WZZahẾֵPz`)+mGlIzyJo% Ϗ0UjxQg6-p_M٫ _8N8OZ|km;`ɑ1\P.5 ͆-}.9ޓv|[cЭ^5m^k^Pjx m ݔ1դ5H eÍf . ˵ D~"#D|s_؜k+gKc6&ZE(# g )ZvKu,ۡ}o/3SC3 'a8wKĥ7XAFjkmh$D5QS`B4}{K۲]'sA?>,sUkzO1zQJ9/Vb.NkAI|7w<ǶΟ4Ubw.{L,c05%ap󨢚WG6xΨjg!ǫr :B fژ]ʠ]00?e_$@^o'*"tn>(K)Ǯ8t|lq9 ƲSFLo^UL紞"t^0i/`̕'3XOJ,-#Ԉ2*εgx/?"|+Nc5Sao4:n%>aaVC!PjFNG.L:%j Z"E]: u^ ?<]l$`C0zBz Fk#;I48_ YpNy3+oPmu5ǥP3qϪgCB6-z*URksMzoj690p`h\p34QE EEJ.«Hyb;܌_Xij'n|?qh =.YXPRu2R.Tw.%=%8s3vqU >ƒd#m0;µU4]׷_<-f!Spmn 7̟̒h3U~QE'uWM f5ZV;ӻpRܠx뢞U"aHcet7O0<.c/<|DG懓z?8ރwa1-m)OV%Ҕs]*b{U - bmBss>~0ؔ;I#A )F5}.Nh7$i}jٽc/?.>{ħ)NA>PfmnϛOo> s%оʒQJ/EILнٛR LD r&$$!my #<;˱~/}q7fBQ+Ks7}ה!eG=!O 9DF&oQ#o4U{J$)dU4K%|B#Y5ịUh @QAsglrD`p'158x<MZg7c[ST,OY&xǛ$ȜCqSXɤRFz{9slhLbmlAZ"w $VJ%a͍J{ɝ7Q2n(lecƌLPZFLBj4BZ"2۝96οq8˯tf/3y&Q[t}$hWٟ[Lr3n麙wzt;lh|[l?Z]/+oM21Mu)_, wj]֭3&a-/ڷUvM(D0yfyHfW'mM.Ktw_-:r%^ Z4g?k}>Lf?Bûw$ 9h=7?oΘ͍]F3&:rRK ԕP2"ti4ƥŒH~.{6{1j] ] bJ*74JkZ<lȂza9].n'ێ7oΣZOq: m)V^q'1! AyZMʪ! b <* ?s~]`DʩR _x+%R.f]QCUYםP!tCe|f/@v*mAVVB >S3u`kx56PQE@[s˥E RF)Bcy)OPh ^&G:"*\p3 %6ڬgނ$1" Y挮pa)kb%&nddSVEF4_Wfȅ_E2Xh5> 'H 0C{g;VN,cT"ƶ0b68yfF=#1>ʭi T@NaF)#: JDV7V!!L K0PXobA$x҄ncN vF.P[,rMk٤P^yQF)Gkg(WpY ]:\%i=)Lhmϋa.8Uf>T(,cxw'0MAp}㉢"Hg~ d4Z}RqXt H}Kle,!]m(L0g^\ su֌Z)5tGBk\(wp)BSϘ8P=t:FY^1Tk%iܿb:nTQBq̆Xކb0J/^ShO0'PcEmxw놾_O5]r_K~i+&R~Nz*%ɘ½>^50rzrn\oMLԚF1Q7l_^"U2@JF/tJÃ&cꐍ 1\#6FB)eocA"w`UBKH*3+J9 N,yg*U+th1juP޺:GJŰE B;CW .]VUz s=]]q:DW?At.Jp ]%K( JPL0]%; t-WmRΐ$eRVw\ʺBW -mRҞΐJu\g0tP.tu>t _\  q0  `T_}n:_V9J6R kpݤJ6_O GWE w1nJ+WT! K- Uq-#w|M YD77^龻"uhS:E%LpT_oôV|]M~PMo\4z<c{Kx? ũ~-gQ%ErTIK&QIZW+~Zyy,~%_+&OT2XHּAՂ_n%Tw! oM,^ıf_ngItmYhZ.Eklʹl'%R:RCZwiQ3|$G-Wtf,UZBuo}3q%>GS21\{D!- P{Еࡗ Sx$@\*AkONP?1&>a$p^üaS")C3;K5nj2E 9Q>`we1ùQ2/|dLJJ-ѢǒCmw1%a|nC{Դ;#3.;DWXfc[Oܥ\ڧbVw"W'>x7jm1܍c/AMVM:IɞUzv'U'*;0/4:0** 2l3 .~xøWrfli|?*I]W=%x W1TARτuDܗg]󪦥zC֜af]rCl7`.+WX_ }h>nkv3Ħ)=B#e5]BՁs<OvÜkb_W,߽iYMŰ YP+R*aD|"`%A[m)%eU6dV exMpQq)NHByrT@NP*U.*- KC$EcjTNOPE#DF'&'3Ez|Sk[ПVh/՚]1^`o14 .`]/;Y5_}u6 z0ҋX ї:Yl.Wnz悜))4$ԋdkxxMD3pcղydi 9mi[)En< u|"5I*<6*+DaRk(c$rq89"0)2hԖhI Jc.+[ H+)#)C<.(-"9e,X| =p2}jYZvB9ArBWP.r;G爟VFj x{U[^F/opCclrR8po0!#C»=Ea0-W'KPs+/SeQE +:NGThjXe_p6Ihڟ׆BX^$YoᗰgsL;_b1YOGL{fW=܁BX܁IY4^ 里o NΘB" JHa$hNvoL`s.p10nt_mGkOӞ>}Pru w6I| e2)`N_?Ҥ_珟ʉkJl=nϯ@UGTHk3u{ޭw`C\ۂݨ,L&8~H "ĭ+˸űEGOT?::` i$$إ9AͣVbY󱜞wfj횳C`GWu66"ڰk+eӶu>ۀ ͠ Wo%fn)X3l$MXo $pf-juUO-Xjmv.-z"H)֛jM6ޙ\FKwVcT*zFҵg^OS[062[ l_Qٮo؈~lŶ5 0̵)`Hڤt7>-.ǟ -HRpۅ!/|Mۣ Xʈh h" ,I( Hp?3W[mDx  nf"DZూk|7Od0IlܾAYxcC)Ņ%IShD!Gۅ8($*P&6FѥT`GR)T>H9kt")($MJDD h|STĵAr9l"yesX4a)bꠠҜ \ě"v(u1㍜LHpo8:wDv~髥r ͋❎*<i6ʜ1iEOFz\khX PoHۜߞr֕K3`lTFHiT̏ d;Ǩ$YO#(9 "#5B{/',3%\F b@,)DтR8*"[ ~k'v#C)@:@gJE((8oH j6';?q:K -7*F=PMъB"ɦ]6G,^-A[Z#}iZ䘠 b4Ÿ\=eZRZMA4Ǚ[ik#PVɱ_c,rц΅oF \4[kNIn! #$ѦSI.N&bP#: %+ptu,*a`ߣt6 kp/+"EUYE?'6䉂Muˇ9&jq׆PRE1BTDxaU#{' Z$Iف'T*'.j  bGH2 w$pe] ikh)Xjcrv#{eZs HZQ @Ҥ F O$',dPtD iWiLk׆*‰d)(=(<= NNqן\XaoJƼPJdteT&Ƴs<rmtnUq-? ߭; ؤ!0cDXQ`L&BqZ*Z]LA"Lb+H !>?!@$uADA9\%#:ZJrH8N >?T{{veNh:s^ѳ}ugp#b1)c?SX>NӋѵЅ*7u='!._6 o+HHkLLn4BY%LF.pKwV7[=%"w/iwIwdf@/fz79ljBѬ.wNMȮIN%NA{+>B잌{8<.|d2u[(N`=)x*/u fr+r[9]TL+l9q~asu@23U*QMۏ9ɴ$ MIdϿ% *jKOP_ׇe7V*VCUYjeXRWUY٧S9ZkUUhMd!%+. Et¼*-wtU$b9_wB9:G!#*w?wzw/~~2/^qާr-$XMo"{U;uK۶/ߺiݚ6z}D6o=;1=AK㌆\ \$bLҪr xhxtGB2C^F InIh":YMlNHT3j\~q6H!:ݜN/ې ّB=`MhQ¡ទy;Q V^G|gIU ^6Y/EGzyGzgp."sʋ9)NB4e6zYPN=m!^ :4G>dIPDRA6#zj1!15C\ OM'kM-t40Ɍw[9stK.lCs}ȗy6pfݸylAmxFwBey6wB!r'Tc*S;;>PA֝EjI ")pI%!wGȁ=lRrۢ{x*osuhӈ#ĂV <E Ԝ2%)'zun1&L|傩DYR9C4DbK6l7FΞ{tcB 駫jQL|K;|m/}(sVwr*cs?Al¯=vΛ(aRb"`޵cٿRȧT@Xdzэ4_v|ڎeɑ8A*=dKvщXX\υrIgcrG B]`5@_p\j278m)i>xN 3vfH =8J$N#q6#|ES DbP9juZ,es9d<+SfT"iy ІsF)&cD4s2/)bV! @Wc,$ET?V"ƥ|$`,Fy O ׍~vyA?gFzC[ƒ0zJ-G,L):zOr9FpwUXs`lWyMPY3hR[ko"f񭲌ӥk.L }bI%U dtlM6}ȼ ?l/!݇*VLn1*oLtͧ79hop}3-&6RksNj vfɾ- T~ZqYϫ4O] ul.˂>eo[G oMgWIO{zMfxKNv2IB9K>jzkUZ-ѳ,N:^^g \xfsD\zCd&F[HQ35$I']¸+6ԁirګL1}<*E\3TlMnEy30PJ1W~'wgd206XZ2FheTk4l;R Y50Z {P)=Ȗ,!QI#Vr쓆(P1J-ȑwԥIɤ^(E h1=|aBI Mӭ2G"` nFRВP]Tnl~JʺLU,ٕ5q'.QN{TCZ -ufwT݌W=Vu6atؚ[M'H֕7&BE[9X_nSe]:~*hζ]v[PpYK%D]-QȽi+T↻,f;ԫgEwϓ-6z쫹w?Eݟ@2dixvΎ͍)Tr8)//2Pv1a9JF.ƸT!U쯕wnDfSX0VߤV(5wKja tU}o7o$@iRs8T6yhSt5l=0y,ogҕpme0RUРΊ_돪o[^jQ0|gz;Nj=^܎Mz"Pvm?Ҷw~t;z3}\Mo|īHڳ1v2p>jFRqXt_J]1Pb%Pi]F+ꓷ7$o-w0|]ٝч4>)H ZGK!zƌƁZ1b"iZ+I%F8 =Tn"ϋP,`9_ *6Q Z}zXL\ }㻢X64r}uwj8d?I.J>&=)8i½KaAzr=sy^buWiak#|L& @ ?'vHFd8.+"uXKq"ŐS soynNWuss;mst d~0XJ9ASDn9&r6ȹQjE[;lVJH?B"+w>d.svB+B4W"EIfWz}k/Э'm}L!\Z`z2 ^Oet}dZpJ |Bp+r:p̥T*Yy*Y)zzp%44'9!J+}2pb$Nup Jj1:!ZRN\FN&.Y;W`%}tJ#W`]%s):JJuJVjK+{CRĞhT W3煫ZKdg@pUW/1 _Z\@\|*V\VkJh Z?dIXgO??-ƣi]4ɂ4%%U(J15z݃#ТN/c3v}0пQ/ W_x-whZZKs>Mn~yU C~7z.0#R*t 3;jx,ue[  b0̇ѦH%>EMnmYhpqp.(tEaB!QIxqF-Z`[nZeJ7Cf%ov*JU*V @[ ѲUF;:EUik 2JBW-MtGW'HW@EtH]e5th_<(tQvIҕFh"BR@k*Õ-tjt(MN$T0cq3WuEWl3Qj ҕbC f-pMk eMAGW'HWi!d+2\)BWn|9TSWHW&(L2\ujh:]e[ݱ9\2X{о{MtV+]юnzŸ^Gh2]Mv/7%X/`Xt%x(Aڣ+\n /l˥(&(AsEbI(Px0p\1[12`%[12W x@kavu4MtU{kg?S+@0`v e[*tQ6l^ *ZDWXpj ]!ZEOW%)ҕ"ʀe{UZo|(i~J LA[4wJhl:]!JhZdGWBW r![DWh{ V4^]ettut%SMtu*eq3ZMRtsWHWFJ4Վo3֨VUFYqY;:%2r2mpz4FF#a&%?E;{u-VJiwޞ-b㓷g6vtW'c]A{ gOm_>igmQ2 [c0N{Zf.{#.iQA.q6 '1]eTwtut*&go\њWuj<]esO*u% 2\ mX'Rt)ҕԊH"Bg*å-trh:]e ::AR֦ X ¥T2ZMi+;z҆&u]e5S9nV:E2FVi>]e=sWbm4>EHGW]y]PbG%=͔b(wwG~pՑ_1XgRb{4hЦgע咴[Ƴ2^ិi!G ̍8-sSٟoߖX/OnO-PҩڪB4Re9#9Y9 %?5U?.,?TyX>(/.7= ? 1C;aS2Md2 }̟{Y֛7 Ң(\hǟ kMfW(-J2,Zm}wp7<" ybmYGݳL<6V 6q7vg-9{=9G'l L]`9u1KtXɖLi6gug{6 1e)S(3s%0oqus~t6Gdѯty[2o)2s}VH^?n">zӦ("ә{i4j'w6lʍ>Zʆ Lnh1_SſK,PΩIt>'}s5fz~ ].rI#l9׆4sz? ،v0s{#=rˠAhW|ɹ?||`e\Y\`cFW#e_zi tvWJJ]le Qk67a+w,p+V{l]0KF/9P) 6\F;(DϤeW2E,٢5Sv~.iNIs\K` hiFN`IiDETa ҹu(I9̓_;~_nv#p`G%;8a$i,#Pb$Z0F2"`֎(ÈJ(7|*GpIkjlj"Qk?rTA1FY睞?˗v:40>tg'S6+j 9TTdeyIߕlhS 6oxRJLLXǥ}"LBLB2*#yWDuH2Bd 6X[J؄$U73g;3֧Uj,P.5s!Epm6+fw|`g ;>+ʣkO $![ '#N)pNVx5OՐ=12mr'e&d>2`d1t͌]FPq1uYǡ 56t)ا,xD́F2pkSe Ŋ*y !Z0 DwCFeHhϣB x"^poD!ģS5ampaԏx1cS68E͌(:F)G)nc$LXl_be P(zu؛b75m~>7({HR%G︩X?Eʃ"0^mVyFC|֧{|X^\ǥIWT6H{v;w$'&/t E VRqAyüLX;Њ4Nt:$W2F߈;ꍋR!Hr!N&&kmG8zTsxFz_q!(\Xύ$!JʘT&H@=uh}b)ǔTq`TӤ C Q~ʳO,ўyמЭ~7P4|,>:y$9Ѭ_3 c;/|>-: M((s0BrδRYi:N~6T~ p LPjS$ 'z"F O꬇/Ɵ)k#1q]6 HsH@'DL68;.;닛oD|}q3A<.vYbd@jhuK$ {;=cl҂ h (ς#ɰ@5\S8S\hd+'+kPvHTT"&j#jk%NR^r ilku7ЏVhiy OYSU6gy$h,ũ,]ʇB;ybB1M58\YRaN-usS*A⋺ ,TAW*\^ =p\3B ޝnkM'6Dth.*`wItOZ#UYIutS!!5Fv-rmZ#̏*6 ~ݻ_!"Ax袭c@MN ˀ%k dKq۝@v:'1 s8>.^u>ݷ`fKqʵ4 mX_bc`XUPO7aqat{r(<#y9$IARa#NO0_ܸL"`, `$NG}Sk7W iyWpXNeu*kU̮]? 㖉`)#)-NՋ)FXiq7 b M. Z^8l!)a FNlu D   Xb7BK]Cd)%aRKH!;& Jq& gRXfAq㨀U(5zʲ}wUɃ*hjLxjWKR֛(2({6k޲Fv&2 yy|Z{*%tF -̀4]EXc!mCѲt5ԤT*'ՆD?'$t&/lHz<ϩD/[J@uRUc 3d{:u ( (N^d,k2B1C\rbϥ: =ː}ȘNxѲfB&ǧą`X@P2*xH i# h]u5rvs_hC W{=:?^XḦMa,}0"%ͳIf)Z2'L y@\ OTG m9ڊi%^78qo_Y 5IX%9+28KNID"J!Xegqǚq0;W;º B(֓vN .BO@<_CAIK )S/a\`l7Ȕs^0όqVl Lb9&]A'槴qVS1qˉ0$&Rv%]e3sf<=hTJR9 zn Fl0}}S U95ONƳrQO~!>,sz9 ]).q%$,ޟswmtD$\ꩍŒIƣ:U&ɵ;{b9oVϥu鑆u@w=,J<XciTTdcz"U QQJNjoMFlHLVmX2I|]Ѓ"zM՜] j1gYHɵZݒ7nxpY4޽w7gߟy 8{Ƿg޿!sܧ)5$8eMo"#j迼u/ߕӧ.7F[qXGi+Ȋ>q28HEh5x K 4'/ 2bX0p !@ó7M>.<%%$!ə,ŝ ;i=ޤK KyɘiRvn}x-% f+Mr<JjE%(2PdXijE4уJ\,3Ɉ%sF99g(|@rցY=ȲC8̒R &:[' e$' 摖30PF $3%mtoFܽxs)5&E( 9qCYir}SmY %vR{ o_ P猗S>af#KθDk 1BA-&eI >[aUf&&浅9dKL*WuC’jى>kև4G{N*CMy2brE601$&aj%ԷȆ2 2a'c$cWivBY~/ kOփ+Epu0 T8Y27Rz[,=O Za|R+,=_PN;%vRyڧmC}‰ZUt" \VʳPD2 F庚]{R'?Kͮ=Cͮ 3kfjR&,F% JuO &z%r5yr֡0*t,,B 5k,?[0]S8jB9R Z+TZv %Kn^ժ~FPa8ҖU +w,"VH=Ol({l(.?(__eqdr$&i)FdfIy.1nj\p߷?J( Q tV)K uHN*a2@hY0u}fe5rv< >VїzSl| asiə;oPc_n7j3W>W6V K"q $KGm0SԤ`Er[dkR h})iP1nCvC#(.zSAB IP ٭T X;kJlЧ%34!K>^TGݭm3o)"jx߬!I\4$A' m4 ?څ-0KFI1֩+V<vg4y^sndH̟`d@$XQ[IiEsANR/9y1)0O "[e(#^Cr`;{=?ZXkdvW áCۈzQԶg e@dKRJphu(~L=^(3&&PJ1/)1W 0d6KPCXAi9A;"ןGCښ{[GIjE,\H`N{WuzP'~_#-tNx5$H֐PRg$)N{5 ܃\mس/ͰHR>i'ݾ_oۭ~9iS׃moIO4 ~95n7~*Y>''{W'Cc_kLJzW%kKǽ9 !etLQ+P<WI2m lyBEA2;eF'I"Qzc> $% $ǃٲ]x`Qޏ(+B藹0ѵfnXy}e>^~v-\Q3+pW yF&Ρ/E/iҥh~@%t]*:A.;6XkN~(![HZH\)F5}Z9@!r]䐜<ѻʓv Khdnm]k|H𧋲1-^7Xh_(BmX~j8 Ū@5`j٫/djgy.$0UV9(/3 Y>UHAVJ Y9tAk  Y݇C3dZ'2d}CV ֛]{o7*G,p>'9 ,8!QdFfG9wbC#Y%6`fWWG2kBQ!:/Xȓ/FS>uRb Zwwp~YۛՋC1yOw\=0EYuٱ(%r-|Hv]35 ܷfS 7 ҩ~uar c׳!_6.6|>?8r|F ijLQ0zHuMiϧP/J; E}kt)]Q$QU;w¸w+Pze;~q(nQl|=ƨ  W\$O-HC ds)`$/Lmj`=5ma€ֻ vKa%ȗRka7q,^-/Z/,}pxd-U74LJnGoy<љ>0oG_^$!ݍZ9XڪSSZN&2$΁ 6 t,>f": 49+S$xx6KxQ岇NKuCǖ1_tŶ w}~di;d1y&7=“XɌ$|MMQFEFĝxW]sIQ7 ^,8㑒6),+_u^>C&R3=:]_*/v56KZ[8E_ORHWԕNh4}Z|y[ǟ{bO?GFF+*Y9}DduH= BZ:~~=v 2nG ^:hʆe[vlgrX4:7hEjNރrq5^$6]`;A$ir4'Ixp=HzcD.ecE!B&IZU&HJRXwS=k斆 Tϐ7g}ď}xd1[ߤ6>zLL4JC:-f{xO>p#`}n ^j9ٛ *(xgVFJ2 OE5'(eYa;+A:=isk+#O{`mVvbM̚ko!RO7ǜ`>P $(^e(Mi"b &PԎ N%8yiJe@ZII/d, 9L:P Kpgl{@/'ebG@-\+zd<|ԶjrK&4J445Ng$]J[DG#I`@%N1'J'F# ggfMDۢ+(V 7R35!F/H*6wE߲Q7Ѻlvgglg _jM.Fcۂ5ᲂݐ5󷶢í[ 1δO۩f։gØ@m}~1injK"ג6mB[k.B56\&2m ӼePaèz\Zm\a-g#EӴKw V7MMľ϶Rթde} JP0{UŦHGe^x-hCIJhI7mAV! ;YFƮ=+T>9I٠gϾ^¤*b}RnjF&0Jh!C$hļk *p^ o,'25ldgUI&B6$:,t\aTs(.ofg|fv~ѝK^T9N{<Xe tAA1M(G6:!lѓbG)uؘԐX6PcꙘcB [mM#iLjϮۦwEfUG֋y>:;]X/^/zq-Ǜ0*mj  Vn"̈́PJ^| 7+UbwӇ:Fy_UT >X :~dYU`s3}hS18uE*b4/OT.XQ(wn<"ub.ؤI:Ť6F/W)/"dREн9\hJd-rZheJijxO. ri䥏WiUvCXX/mg֤v).PP;zFgMd^irh)PQK $m^ E6I) 5ZaA-EmRTgl'SĜ]2Bv[.]X$j;I!9煶))6A`A"63rV lp 姐"r!P%+9)s.(2}juUL^?S[g* ;Ir5Yd75#(˯;27:3d'-wPD5'm ,fA 5t*(H:CSo (FnW(1M.ذ3ؐup1ރL64y=sLr 単F:G۴bpZ/ <"L%jQEv>ٱ5{FF%eEkӥ1A fLg2)Y#; ;1W̝qȮҡ:-:AucMa:oi[<mkH8y@BlL&[}yV٢ۿvRJ67IҀKEl{g (՟}ʊfOQG-a#;@X&Ô&x` I cfca'IDx@G b@ ,fe^c;#g˒X mO.Vdx-F W|y`K_zЈVY,1sSrl2ʌj7TI6, A o<4"NwyEZOrsF3w$WJPЃI0bJʫn׽}< Ō{L#T0&a/u&ף0{(]]?Q#2vԚz_RO5mpGl@()f%,!f!18cea- jA-zgxD( ɐ G*Y=ᲗRDQc,*/nNN|7lFiu3`P>44mnm)x/R?a躭F>9RAjߧ/эwz$ G;S&98o=Q!-bؖN֎M0O-ʩwQM8c ?E%><`1!_^BO[Za=Je!sopJ =nj:ͫ?b@՜>;(s߳kDuk9z8xx|դ,o; Ī^9g:[5Z= ,3[LvXz9ON'ە*~Cu3+hmo^^:lȺŔUI /l{ +㚫KsSfϿ_J՘oTt(X_OynEjiA];\z8Zz~պ)bwZ*$eGx=8lϷ)UάyT˒#℅O~|O?֯o_o£7Oٜ %8V=< ೩S:>~tGGxH Qˠ"&g0\J( k`S K5{GÓ;򿇙&H=y5G4Imn6eJu_T[bWr+?%[[QLR]߯13H(r(eYbf@t_Ld! AQפDP!NVm$QVs-J\^799z{Ύ:K|粳Kqٹ恧+r4R)uїz&Q^O+!"s@ExV" *8#"MTT,%N9J 0q "hݵ I$5*Xi˵$@%I'F(aJl}FQF{ˁWnpZ Dˤ>_Syne!5I*ԞCҭ2Pt|b:XH VrD¼,9`2Kl Bo2@Mr!ۤ!n[fVƠs"e2/φRIjȵQTZgL)G\Yz%02jNjk~ity5 hZ41iYn\% P.7yV{lķSɂF B{ bEk$R9-2j~@.˩kMQߗ1OWZ+˸ͱCG{BGTڃBG~|̃ޗ2DL Dd#3o3MC37yw#^ލ~p(zxpe+%;'!2O(seݎZAwKI4R;D'.uD G[  jU93 Mf!7#ni>JsЦ"[ݶ\lXL*p+dC#uVX_~W9Ϛo~M_0E]܌Z ?G*o@:|'32j\-/Ûь#!`Ǔo>pL;;L}̹Au D"+snie0gpJUZ+|#fu\|M.ADv8+qϿjO郦v #/ёb3 F.8R逐;:,_ MUy_L[n3ۮN7frƙ}PszPѭrEuy'g OY6sK*T2bKQ_ZHV߯eu(I8u #]m$;%tRN^-oG[[1IQn(Sx9s%BI؁aDDbe 2`]NJ-_ZHr5 I;QQ:*-wIH\c"@!+&0$L"5c.7J!p.Z#iGaeap#0#ko e6zi1M{6Jpd &0-p l +[7zo޾硦7Ί|MK9-f =7ݽuzg\>K?-h5K/$ {j6j/,Gzj.9p)E Qrci  !r򞶛KIn]٭&AUI !#zj1!15Ǵ҄Qk< KAH3f|6BZ٥G;^3J+1 ;kQ~K~O7h}snS_QXǢ4)(]ԒDk]zvCkCb96rh=x}~tw{~ ǿB3eR@#@y<E (Ԝ26K56CLL0A)ML DLqЖ eޒUx.HP4%b[ƿf↑w ?6OG7ըw| ח\vt:>0F/wT`1,R 0Yr$i,"#H% (SyaN0ta/4F-B`O6 9XXǬc@2ptg!nj4G+(o/s*zW4rm'( J 1eL(f\DK*elFp4N=$9')Uh4 C:2\|,ׄ<%@'yZ<|~/|A%ѯz 1'\h.N4(:1|1I@z@uUbAxZAEu~I!WQ'-1P{ ]8t po8 W9VDE4)_qTEP-*7~OC(& %d!9uQCZʬXJ~(Y%WuHM uU2yUhfoN}c)/kxWgKӱ#de$K$L}\]O5˹,\~ ٱ8Ə|h|<6ձgJ{MAL^IRI7‚VJP~flR"|Vy)%BHiAXK# Ka'ЀF## -YƒN r+Ҝ%\Ղ{i .`$xȎZ .N|eYܬ:%_fp!>Lly4h\Fp]SXT|Hu5> +gkubu69~\ot=鮈l]w׳Ӎ\#J,ⲴyɸDZd4՗TOeTՍK>/.➒qy~\W{guz#@g~2څ+b귫Ŷ`\(Gu.&28$< M]AhK#g+S(5q Nُ cJw+,8;LݚUн[3Vwޭл2Cw͠+$q*~,*Sk١+SщF\5U5Py_ykgqLAtXPtI2RKwK&m\*+tZ2<"BuųS}槿|?ۍ٩+i6Ӄb.ɶ1Q5iѭ@?՜ ߿ _ Ԓɀz*?Mlq{ss-0gpߜ|&/o=KTqN(mQ}KĆTo= !Эτ" yU,-EwRE^_S:,)&&|K1}jA}z+x#Q}?g}@KA5>%)2UQǐ4Ւ({k]UM因DrֳL=BŻƧ9MmLS! qZv'5X([?Z ]J.ZnF1v4 x:L 2`H҆'*8A=pM]D/?]E!JRZ/R0@{eAFvAœl.&XCDUO^z'*2,EǷ2u-;&Lm钵;jݽR*VT.Ts0gGg29($% zVIkQr &SRDeycd뚪0+m$sB&dȁ(ѐIȢ(l؍RVE2x_YV#g>leE#V}5U{ Lv Fzpɹ,J(Qf$2j F|0up(x~3 m5˚J2P )D &hc$KZm8ĒZ_-rk/H:^6 :$_g5.W/zQz׋9^&3_mr:D$Vnwd=Ym0CcŃчc_}h+C>| {5Y{Fnfw{Лj@7րGW"oG?>ûP'r9JjaoO㒭l.%b(g1JWFkt ;2 Is LZI笑{XtxKsZE~Z>Qj|hRN4 :h\9+ЙcGO&>{zg˸.}[y=<۬vIOf1ףrq\KA9t@[zN=S̰JP/!}0&5@ 8.:uݑd 7'G\dE2U q?dV !jk)e/Zg S`bRt|'ERs Jr>Jq!Q:eƹ \g@ѭ#Ykc!'6`TnR }HI,lw-W#ԔZOԷ~q;Q(y=|&)JbK\jOiuXpg[_㸆RE0 Hs<5ZJak~@"Ԗ,tFɑqwA%K IκbPN?395\DI'󂈻n,M6;vf9;YyOn?!.Ά8-)Fṿu7Ҋ"6Ji]rcq?B$K3 h z6%֕0}ynې掠G!Y2y )f6!fSf :r֎Gio>  ΚߓAq5RЀwq>F0=*dէ^|QVv!v=躝&i4xak3!-˭n<{|)T[Hۍe6^6A h΄Щԃ$5AkeC s %-6:3tP/rHKK: cP3Q7Rǔ`fbD`2F EH>37+sث=tRW(}8 Ͳ ˶+|+f(؝gO#1$Ft\0@0LY%!Iΐ& d蓋~Hd Vzdwi:R+J6Va܃o{y8/InSCIj2fe4 $.cHdtɕ]/E? .EPRIOuXC0^GU1DX!y c*䅕.1#tZ*c&1tn=dږbf. 1'vRk#9пX;9;4&z@8nMqkgt 3vV+cCgn1k3}*or !nMJNE{^?Q9WbimvFd ٪@뜐,g5&R^rInC8; Xa H>)F[*De%OJ <T!%u=A2'FFnd^.Ec:FA0(wU#gG9kw f-NN61*]~ YکSRBʨQ349'0-M)%JЋ@M|HAO?3; '@p2F(U)09锝t$)ɂULPԊQo'vJx?w8]DٓȒŗV\:rc7֑'`xYnWUgƲOƒtr5߃ v1N9GM &SLZ2FǴR[o#XHę0fzi!./)[AOEȾ+ _LZ8x|G YV&4aKW #ĞK*e=J|dL'lB\$8[OT d!srZXvMٓr3o{sZ-͐)ܓLJ&Y@YGSd2m9:kNhrJP)jdyG7IQ%7V&CTt֐FDE-Uz{4A21 q^F]?X/V]1 %/K.dA:^6I=?OƱ:vr[t٤``q8hmgƸFi :@D/*vF e>B fWU<3qδ'AR뷩,Kq" ^\&2cp|2׭VAf:Gje9V㓹.- ]}>Q,=h ##ydtm*NI3I*ʽ{Yb[s9k32#ݞf Q;ZU!'8L5K1>1/<`ǏePQ sOhgY߰J1qg~cCgzfyW,qͫx;8E wٝݜo 2h:sGY\[~]"9X\tE6*g-pni[vo$ VW#oFwE[KGB]\-?]q|J415Zm̻5-FË~m+?^_Nx!qK̙}8Σ|o5`<r9k5bpg`NZv w:뺍辍uYg$ Ǔ,XdXeu<>@[/z}Syˍ:jK޵5#nh~qU&=lj./g*@,HrmŊ$ mI1'$Av7B,Péc{⃖4rh%U)8&"BÃ/4#M>*0Xod1:MdeAP5)R4&6'$Fr5ie?\^ou]wYtvY;0,ly-9x+JH.BPWZ}ig:EE9rYde!ϦM`AgD\qĉU2G)TU5r8ٟzrVc'O0'x"O?F|tIh:ܶy>,OgCxmQJ74nѯu}YT?k_ f/}bT]{VgWE]ԭwr"Cs'rcq5IUz9?_ss4vҡWD ?`NJf /#γ$U ^΂قg>wE`m=Xp$3\sʋ9L(4-eA9]W)@>#GUI$% 5ƄVvQȹق[Bc͙.mx9AP'1~[L [0h[? vڞ,5͹CAf5?o? cQ[z. YFjI5=ᦻ#Ly07{07s4)>A57pV>gLMQ{s>1h ͔EK\F4DQ4RsH,QD M*΍wL0A)ML DLqЖ e9ȉx8HЖ4%b[ƿsKߐwI35ꝵ1: [yƯM!0n ,erF M̔ | afD@Z 4L#d^>bxL/4F-B`O6g y.YK Yk#&ƀe( tL(05Lp暹y="B,9A]h{5rnw`{6 n~ nidiL^-l >kx܃8J4rm'( j 91uL(e\DK*ely+x:ճQt[<j#`TiN|Ҁk"!PXN;-yJ8Ny2\wrD)=kz N]g`9qg]X iR!)ˈ%bt Zyw{NMg%ytDlZM;Uި\V.k_>ғ7C(iQ<[C1+ Vt:S~䏄Zdv}qjӋ~ٗn1B dOvUi鎧 O\ryrcnآLn0zUecg\f ӪB>'LSEOMf4WVo3d:2k+,nzbӬ[r\=?6rBV ^x_s?usr;Ͻ$lv%UW%~wY_:#}WKTN8}[o,eRI8]9cQ55W<1=l~qvn[R!iV(=xx9Z:5#aW d!+Tg١k?rQ]xwcEMo5̞W/śp: ޼Zz99pLNO罺-euwV凾_bKϟc*kaOϫy2gYvzu9EK7~5Ws/U^z3$%'?W(F|ݠ>˂Gsk=n)+_ Lp*Q ӦsLazN/cA-Bj?=g}-1s<%l`Ѽ_ߦ߆ݡ߻uiOp-1:-(MP/w=IMi:fsoY,tl6m3{ ~iix$A,ۆjZmv<1I&fR="ՖVV3ﴻhc$u\FtS7<5n.VUJpZMy={$$*'3(3ޛ-g;90T9т#P<^- *!&RcqzDM4,d-GM#90Uok7vv=뫷t2.q 3 Jc+P:D) K|111T6T6Sh,C:Z'],1\81Hmc1Hg<jUz%8A8.>pJF"(ZsدJ/ic]hՖRyv#- ȶ%,'cpyCN b:'j5c䡒5‚VJP~flR"|Vy)%BHiAXK,wO@N,G١ٝF잀NdKF;ZRȉJsrS{i yI<-tT@V^*|λe>>Yu Cʘ9"%Z;Vi(FPu)I,_x4NeVdpC0?"VEs˾v]cjb=~y.^ 6\*MB%IR}xKuKV4U!w^#(fuݪ%nI28)_N6jF~?+|~Vl| xυrT"a(2dhqírq|)xPmp95bQSk%w΄ٓ #jɉGSm7kɱTEj)Շ^m7SYWmVͻGTDȣQW\E]ej&u[zuo8YΣȖ}R29cN"<½qX8  99{#K)VK1rF=g"");ׯɶG#6F=6xF9Q,#| ʬj6Q'K+cFf 31f $7R@\c3&lywK#[) O%̭9 9|_0 mFi.f^ ՐR"@""UN{9Zj=N{b>`y|"cIəiO2}Y[cXc@P^jZcC q"EaQKv"og0 Lȗ *}`Sr`)VgbgwPQy@m>1h-ď&]cH(W %Jx XPCć% 6 ±64>VDIU@0yufG^[xُUAjL@RZ7faVFJr߲k&->{ J Z+{S\D?5qܽXe%C fC2a lB_kYRT?-v UP;$LXpWL>A@&#w(!Ȯb@j;!7CAJ.;x@LAA[N@vMB(#2Fidä9lEOMWQ"t/XYHXu0wL@FbkY*)SbAheU\7Md& /-H2:@M=5@Q"Ł.0͑EUPjϒ (Ez@?PY* <磌Pc4i/uj{UQ@};ZEΊ 1H5ABR [JDFࠄnMYֻ"1 ҈9>4TMiB_ȠL!rL7~CM/Vĥ&ˊ٤b,FyDDa:YB12燎 LtgUȮ;wb<.0B̔ZI|t(#KyD n:@GvY,gg*ֺTYQd~Q|Ѫ`QBWE`Z HDNfʇ)UGc `$`)bY(^RX"ɚl>A& qm~)Jrj@#;3PFFbޱ"xa;PMuh 1HeO{?ѼKȓm1f,M&XuOXk tdDrHc.m,56X!/VsP6D2债 +& 8'O5Ҙ j`UWۖc)h BV uPt ˈ)+>$-e^U3rA`X?6J?HÙPȢ td!nd`mBgU酬iFJkh@2~ȃ!:8vGyPQ`QÍ'piUR"찈Jfcr"cq¿? VVHWYtMSLDHYP*VDx(m@;M譛ޫUDK蟂̄j; HG}>;>'o%KP0#ΒKF+7%@eŸ Ia = B#xBdK`3^o#G w`TS˿B{]،4QGy[Ks򬵤llyHYS.n^PAY^ʳW=vhp}Uewf":U'%\+1}C;8B l+ Rk"K3HρOlr<XFշn֓bCcB|?y!K}ؒsP' 7B0#寺țbCw F-TyB᜔E 5FHOub$71`bRuM` tAq1յNlF54<[ WjQc %ߒMr T pEC!~]Z߳[uנӥd("޸3 1ѫ;ȚQiW`a2҂N3 _̂B\Y<2čtXvo*gsSj̖^j:i\@C,Vc54&GH*8ėԅ˯q@BĆr^TGB#R0bAA~lż݋[//9ŕC@y\ ]dh[{Qw^ d9jI-K.<EF#@?ۛ-/;S^U<==| czN a`6_۹h9z't:E'ƫH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': tN dBmf@@;d#@>@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': Nvy"ڌh 1@PbYR': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N ul(p0v;N N96ģwhhV'):b6H@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': t:N#N>>![Mջ? Wo`Xْq)h;%M1.EY7:zPU)CnO2N|J˛+ Kr\H???y>}}ٯwm[~ķ%]{]w?̥^Zfԑ'2l6 Ft<ۥ >㜒7nKRfYfnZz9Aj*=En$K+= ] \BWͱӕ*%OBL!`S ;RJptzdiU-b9i3tc+AI 4-8o@4+n3t%hQSJWo7pA2]{}xb"ӝgAwv"a9lzʋxξf溅/n68:33࿳?|{ sX{fLxL8W]]^/ͽחzџ @w``vvy!/6@a̯._Gw}r+KJ4q^~/GzjY~: Wڳ2޾yѮ^.wЂ?}Ԃn>Gd7g^ aoxv~swJ+F<ɉ ˰4j&/k^}j!q#Cf7L:2Cs%Ֆ{ K ܇_Op[yX a(G!{7}Ч/OWo7\~7 ]{ՓN~ tϱa}1<⽯ \\2{sDJN!dxEpVܱgUĚUO0z9m8Z>nm4NW2+]"]7DWȕJͨ+Aӕ̺fqtŖoi+ky+t%h}v+]n*sLo=prMW}'G&zc.@WIꋇȬ(x=sgv7=si=z0 ޗczH_رR}$"!8E"e,!EoZ)Kszzg}=UݘAr,DfoQyP=n¦S_̠gb6gO׻QOFw6=yW8 =~|BX\b*:" F*2T_燞%E=V/0ᲊWyzoo)kv66#J4^g{;*Ji\ZM+sm1cKUpt^H( y,E-0 }cq^2?:./?޳ )b cGؽvpkA?Yߧ ח˵5;NN.NƳ4إ3ZSAq.J\HC\p< N\tL >F P+BͅS9Z^p?vGԇʤtTzwzdr5&?h/?s) 'ٸt4{I^[|suvU1w =[):F6K 4v[ڹĖs0b7gxU mMuϘAsAj hWqkU -&@lUEp"ZY dI>qh@v;B@v/YY!5}P(LG9ӁS+ȃˊ ։0Aq* 䡤 )Kn}Le4N )DZET dȶz14W׻\h R>TBjXrZyγ+g;sjYh1qf eDﴉL@V#5s0#9UUR"%pЇB!R(F<7+HWDF$]e#&v#c3Cuu,`t֎m!՝M]tb:k)2gi)no]gW0W.;dtLՙjTr`NL_"_Zd136G$Qn<P=+棼/QfUyn+6~hbo=_}˽8LVE;>:kt.Y%h&'LL@tBk^I+BoH>E2]@kjOr|djRIN߷my# V2I`sDE/Ҿ/R:eFG_weT9f%X\ċq\RC58o,'ֳngu%.וJ`"!2='o),MtF`IKD<7jXշ&KZea.} Q2љe:d6^oF}Hi_F/ȹ2vNVJ08UB.p N2[P0ssM]H*[I/"!8a6T6jl7ШfD,: ëf\U۽EYcZ8Ϸ8Td"Ao{( &ao+)[?Z̝trҒ:kmc[z*c8W5OrU=AJDe3rvS ^Ɉ :$_b3qΌCH0ƅ 1fYYB[[j]JɌy1E3IQs-sIajf ee_(:_p.Q 5pӃm*\{ݎa ~K^c`Lִ;>Ic2cdIsÕW3 R R2%Z걥6Q0齐MM5^X\sHc+{jl~8 `v-kCe &eytn$$p (4"(˽ #kì!.Ht!2@,:dĂIؠ 8 HŬ" DA>78Cc_*{Dy#nx91oUH Β hB2[%A " 7鬽K]rJ׉wY@J3>`!Id!@r[Y{ Zl'8nfuHsմd_h*E/nxUA3,Cir&*[{a& 9{ŗkiǾVv?.lf$]1rlQ?*\۳wH뒵(Y:`s,C2֯!WdC|}Z} *>[Ao;c(Ĥ옘T!2Z[xEl:g h1tl[[\Ukeq Џ?,Q ^L]bj؇&}=`x;XW=ո+Xl~=1C*`v@:P`oׁs8W8U5Qt_!wPmdR3`<]O79 TqegH" DH,@tYE:׭ޭ~+σIkބ4L_0#bLR&ԜBs麯R\H0H)78 Y(tJaa$WZl&%,yw{OeVp~qSjK_{0|K+,Nj2 ALnk1SxЁNGry11bÅ7 Hs<6ZJag?;yS ezer`sbHe!$@ȃ%3DE8nwsgYsjԻ t`YuwhyO l.x<Ҵh pADϲEtr0RYی ~xeU G pM2qL&Zt' !TɨȮ&v ^ށI^ }uqSzjN1WY+:fF̒\/Jחob\9B 067֢i7(EϺ3GO0 >d<| c[I+w .l$D#ufjp쬱zyȏunugȳ0Y޽zBnm\+,@hsץ ;@bП [ n_Z7֛9otf3vH-jy_{no|=/ܬI1;-`dηt<.E/Ӷ0 Vn'dzSIhۼfyߤen7#nqb5$t{a3"(F) 4U?W`}E Hg4Z*9:>JH-0c(I#Jk9 |8,}2 [ElDC:&&\"coWFNJ;HFJk@*Oi|N 8Jz eH)}t0l9`:Jqi~VW:kQ@㺋/@b[!-'1VI٭+űVY}@[eRuUUJ}9bCNW܌}a $2mH(/63a[r*v9pHitz[7` s %x-689%}(G/IJN: cfn 71IE< XAFcP$Z̈y EFjװWg{*J {ݬI_40Dh%z,=i|9cqM^ IRH f0e- ɢ;CF%dY/:B {]}ѴIKkZl%sl,إ٘y` Yߐbr)de4 $qQ@m4J./W"?@\8Q3n?Zj[K,A "2}DtdY)*J^ZE]1M^vnƒ6e131' 'j#9ПP8;lM>p̛(rZX)/pjY$٬+8~%6ܗCmPj*xB4f/qFh&}ʓ:'#&N:x+#,6jF]仳NUICeQ['Ԏ1;SLa8)rJ'ǤAfMI2tUgG;k;0VQ6)*~ [iGkV͜#F A‡,%QEL 3woΡ"`j"$UY`8cvS,R=*:U jIIOɯ{ήXnB YdIe+ooY.;rc"7֑O`x179*c&We5kKY6#]trʹwOm_ٺ_Nr*Uׯ$rNRSCEtVMH ¬Kv8& L1nGO`+493cb !<"@tJڲF`T Hˑ[ .HUIrdK@$;aP]8U:XtZ\s)hRMw?,!lH`< 6y`5[69aLT_|RJQ&mh̥!>^iޑkv'N<$KPLڄ&0"*s$WḆ6IPp\J1аp{c80 L*&&Gmi\>z-r @]rg{=XJfpSРS2 `a.4֯`i4A`+Wx(CZqi_Ab`Z%`kô:kCX!50D BͿVk=N42Y,NX/6c\ɈGU[(23 <'6R,yqؿ:jkn:zQ:f,Uk#*jfP<ʋd+NZTVK?$*'9QD􉰔-DpDGK)QO  "%OT]E -|421fx\ʟXVchgabv( 709X g #(7 Y?{{/0H0].fjC?cѰD$EmfuZQb0<]觽~9Zftr&s!4NwyNBB5r!} =)=pg~C-˵]AF~M#)q$!׏tiyc7,@hc mVW>_j9f7nrVmGOrӨxiTZ>p `BbPR3Eq^_]4AJ{8$}Zai^.ˮT,zaA]-///M x |~UaU`P@!-+ο,,sN*BխaFLU$1bo0 ?޾~[ѿ}Gׯ^}%3>Ism $U\׿lz[<}嵶эi7G~gn],Péc{⃖"pi2IJ29,s  h4YDxȫtc$7IK%LtXLHT3j\wynH]Nu|xΥIc yN(W˦ ]4\1Vr)uɗz&I̢ VB7 ١,2 tYPF8D%r*aD|"aVVls $%hW\[/, 2DJjJFJܼ p!;V*##g1$ˤ/Y)n< oeBd ^D;*H');>PtP<1Usꂊ:ؘ ҂&lGL"{8&$X_>[^vUITHIm|&>)ϼa\G=Q:;-Kev2w=$w] y2 >%KY{b-$Dk=7a8O >h8N;dFZx;:w#&D[99t_~|6ROڨ9ϵI\ZgL)JHadҜNWqD2E!&.Y#ZFH1"L͚e@Fjk,u #xާ)uՐ5uKd8~>WKcve46nvHڑ=vM}ҁZI *Q҉`U ?/#γdXv*ѝU"q@胨Dz蚙݃JddȜ"sbIFLzɍ)e6zYPN=m7 /8 u),G(YPEVBHZcL 8Ǵ҄綫F*ehV6NډpcI]cU'`$-h&ة>hJ4yBe݁`ؓ)rTʺ#zYwRueݟcYwa[jOre:?Y6TH)LBZ"њqh{GtH߃^@ܛAD4gyxƼ9>mZ-mSQ^L7< )҄iXh/462bc؅`䠘PJ9$dT3xr;=|kC6q*͉4[@!bcvZjp;iyP]w|zyu^$ɤ\sϲjBP3KCV+rRO2FwF{RWvHkaZlm6Tn"VJQۖ]L 4-Km8 ph}}jNd"V, "(oSL>[R̜6p:`0~<= yyO'ɧX~tgjJO:󰑌j+| vc:aAŇ~q#lF'cؐ߇SyOt֜WuYG'E*u ˬG7^zhG_v2E\YodGoE߽ل/T{{io9|Vw7 Wz鯰ï~ _{&3+ޞ&_ŨjfYvrA N'N{|qd:_͙_/Ǝ~ۢKJ{~ڕ8O@ a>A=ʂvJN=Nͪ J}ʥ  {J"<;Sr@6Ui~߾-%>:Lߎ`0};r~\摴[QI -i]=/pt~Yh5j"/|E7\P]oK_'F6 ]7]GgzH_[RfFe`zz~<F)Y<,ɢHS)˰e,Vq\&ݹ[nMީ-u5nO}< J ټ?rWW7L-pn~2uoΟ/n+劉>ݞ蹤hAv˛oqt>\{k~mZo;1.䷧'/C n \ pm&c b2ulKRIS.x);'QY͈dQMvh&c5&A Йec{bNJ/)zP-f+"w0Qdomoهd,?F) B%_w%]vZ&J ! уpMPX7Qײƣ$ad& 48E 3jlgTu:}NތN{Bxsa;; URy{M -J񬓩!"(V60! _J馒8[r.12wVGV,z(GeW KHAJj-MgdlUf,ƲP ыV󫊌7TY|zi_\Ѕxa|6;&+2X3*'DTp%![H+u6 d/JlpP{!+1*PUl`4l*m zSk,MgĎIfbnEJml,qb )xSRKZ銕0XTNFWPym*6mA6-dFيN@eM&"aHnd "QM6tި_4P$b3W"Qq6&[ugwFi&(N*6j*ʒ0SLdSXdۘjH_ lT9sLTLRbKZ!JL fDl6),gȥCu6}m, x/ke)l r_MaMT}5zr!c&8ǐ#[QǾ5n?y"ax8 (Ѡٗl:U6W_Z>XjߝA4.#$|w(+-!h݂vcow[tt5u^o#aެW\e4p=v<5 Z%)mqdȈK3fb.Ej,,i4x,~mޅQot-s?HjGo&{[ozgds'F֘֘61]^WcUlI7xFQ=0yw6r $} *.He{A=kj_"Ofay0&u p\*»l wm`TP$6uzq?dV #,?8B)#%ڔ)K(!L"21S")4HaRʢhm`1N2JKEhb)x^ZMg%twCB(?U=)ڳN{t<;"~# e^ /1b:/lK)eB2̧Ƿ_VDs'U{;R9~(Aekfe mIBB2TP,B fu`xAp5T@)e ) (8 0IgT}C..?}'5$\vno2];t$R].V[۷R.Ά)-k4}(P*L*GQ"N ʂvDt%@,ʦ%ad I N_"X\RP*LbuDvl;qwǵZ`A~`=rMWlW Y HsV M:]Tϥ[E1ޭgi *~nJg+~.bd^MQ. ]4i {H85g^GZ=(n;{2gЊTm 4hm=4HpJbsi[K9) @$KlːL:JIZj5mt0V:(s/o϶}IUo=?mtc~yf۝Q ]ֈfV8sOc)~ů  iQH>rDΞoA3u]H뉐O[#w rs¸)@AH,F^l '{O{'^i6jtS+aQ;c(f[Bm`de-JL-e :ւ<PF`pYXPNJ@R,f~2rzHɃ8ћЂu[Ul~#pլf/.>r3"A`rD^*$Vh6b[f H!$,9ۄy;9_?%[0aNovoȾRM@J5pwњ& >(aȹ,,BH60lMD$#e}|~HE!HBQ5\%C*^JaO4 $n ՝~eNOȺ|=R NNF0z*HN T>Q[ԗ_(ڡ'$*p9J_gHH96]8ҦqA$o.'.Rjz[u}eu펺ޑ;ƫq^8ZUGWU!'4M?ݩ%<GYzbHoEU$.B|={g(AcxF n/j:̋out-/vsvT׷if<IV{d{t!NlFqh6g=s~Dc-?[̼%(EJ1&jutt͟xқ%弩JyOҼ^ߧ!SJcu }^gwjI=jW.SLܟjOh_niyhqy`˞W;&b=lRg'yoF䗫y/Y^17wL?F2q$wtiy7sX~e4jwbKjg<;?߷8o6iZzkT0^:heY!!7^;W K+Ci&ϿN/SՌoUt,X^/e)~ުR;Kx=I,4Osޝ_O>*6*L*+Š7|l6J[|.x6Y|TemG)Apb追eЛ/O]kmú>[ x޵q$ۿBSX~q{cv 6#a[1!EI|TSY4tꮪ0 XQlY'2J4P%}ZDj- w0f,"}: SLi!NU N9=kbİNd8"D:$S, Vd9Vk9񝄼rsXtC:6M@CA*VX[C՛A~N/(4O:gJ; l03lz sZTъd+k,P# gΜ];Ewٵsvy9QybeRc 3 6v!s*> %nN(V'9M x*Ld X#瑔boP}G wկ&.Is7 jQϹ\xOA2 ΏA~׃I"48؃,1Z:hD ԌP: V:"c*5l}X,B0#BXT$z)#"bETUrFxeLD:M䎭q_E;p?1ͬ1y줟f[niʻ7ʮ'8uPl+b4&ƧtUsmt AYXk2D#B]c* p_p\j2`78mԩ|I&><3Dk R {PBN|h-3kM5 ?MwoV$nn͞۴k$֪[•)[*FjZril0#|A,ȓuiXu* 89sb SjI<|:8<7eIooY.uk_M^go_Mrz6=Tef1My Z}6KO̧̻έgտֱ_B}zqK,2h<5KkI,'SO͊Rbrls9A F)z`RPW:2Ϛ(if? `x+ϭ[Ӵ),-z0o^ҋ"5ʳ$iY:X!)xX\.j>$ױ^b_Cً͘/glq?ĪK<_ߨkT^][z G5Oܼ8gӟ^r.}5GP* dg:tкˡ$j]݆-gO:g V ,f1!fI}00h~a}o;y6$9v`Ӯ;LWpyi:cepV0J8FP5D(#ECB&xJVSeұ&E6e:|a8wKĥ7XAFjkmh$D5QS@B4%3⇪cNSn+mfXB<)WP}?$S2=Zjq'udn8"JJic*汫 #3<ڒl]J3UH]rdNpٱA))eҸ@M(X.sh9NhX)Ql=N;[B7zf]CJf/hlf&5.wɵׅ߱Լ٧) օTZr]'$' p9Le6Z/1^'{Ub]U jJ2 xC+IecV1^EŹLJU=u*&]3dw=;ɒߐ%$*iuJ}#<8U!@;9:!L:/59RC-op~[ٌˋՃ1K O0XaOZH/A8Xzq$ucbȂsˁKjyF2?]ԅ˹f*E\ V .(=ۉo'"vn*JJ.O4#H DA-=qg-%В8 km5p k$,Uvl,XfdnCN56̓u{\?}\GmC8ft6N=/EٷL^A߫쭋ç=l|?}\ڕūH’YX2 Kfa,,%dTj,GMP5BXKqV Ucj,P UcABX Ucj,Ty Ucj,TP5BX Ucj,TP5BX Ucj,TP5BX Ucj,T嘯P5BX Ucj,TP5ƒѽnF$sBX Ucj,TP5v^ST![ Ucj,59 Ucj,TP5BX;J.(TS-TP5BX Ucj,Ta0ZH4T7W 7ܤ("'#5&࿸9*Xue4ƕ°󀉲x2 scu uM bJ*nF#XzutV P#ļD&6T<ME) S띱VcQ"^ˈi) Ddޮ& }ݥ23>q$2!H)c8AQ )N pS\a 9N]`DʩR ^x+%R.fgL֚Wntvޕl'5 Q_0'uჶ[nPߩŅe&+*o8 K1H5!-͖SoTlZ'AB0 :0!y#h-hE@'.}ԖU`̙R18ۑ1 ͌]fBRp%sd .iؒyOt0Y?M σB5b['1 ] ,@"FEm K-1pJYj#W42OP` ^&G:"*7y.q# 5;vEmYAbn1" Y挮g5G 722)q\a"#+aB/",Cv4H$!`H=rTiLf<&vd1.Y`pؙ~ h{,K$;_#n[۶`j*XeoVO/jtl2_ 7uFވ۶gBcyo8.ͦ,g^*mw^eH_:4KZ~vMal(&D)Vw,t̺%;A{R_fwǗAsvi0צfųȔJEdiKt%g2jȿ G0;t  1eH*%WkF ܪ cv^E[_m:[WoθRoʃ=u&GÅw~6Y?_Mzu;^uAKk~YБ` FD?7꒲"*T!g"Q]Ḧ%Xk@CjZ+tN`?9& ڢYB+IPH8j:RDIau3n(D| OgGI|v*o^)7lqlUY' *|LiQCRAR9U:%Ir X OtaWN^1oHs4$oRK$ &tN)T,vt[q7dztQ6^- ]a˸nb,_HtB䳚9BJ% pX3H.h5^zAsy12LʇX,(sXI􁬑Hozٙ&^7OPcE]k='zT-m {^w_I94S?ќ z=Ib몆[7ܺ]m,[=od~O6wk2xnrˆ[m]Eϗ|bF5oC[%Bƞx\ZاʪBڐzF;-iyN}9j5A7!bHQ8q4Ƃbv:.=hcSȍw!`;?h?dIK4-6EH4I@'!,:?}f=O21" Ag@H@pk\9F4( @*+o٫>Rh"ul xKDIG}S iW5mr;n]7.>dzWz!%0%R ÕTʻHk{-y'RY:gDa"u+Qkg6c!afbt B9ɃnJ҃ݞ4Q\w@KE!1` r|[K1tJhfY A)gpTDQO0;heMPŢ4@9iUȐam:LMw*5w/LΛu-V]h|>su<_ZLnK%PjD' *P0, c ZBI69mq`^wl~ ˰oiivAds ^܂os c4:2\7T JH>X  (ͺC)[w.EPJAu#p&L)&E1%ČAY8ecET-p%˒ @R6T-df>F/W)/;O; ݥ|&v#rP{c*Uewl(Q(uh=YcAqWڳ&RfNQ8?m]7 I):pZSš$d j-j"0T61gZ*`uy ɄLF2^.Br mSR A`A"6lg-_}ج%]ʏw›FkȅBBLPJ!\2 eb|j/Ld]xrJ!N\ksV%"mja*kИf 뎬T 3@ޠVZ=iГ^IE.QhK`aQ6(QsY:<8b-1W&v# xc)&A̪F:GǴbpLZ/ ~&BfV~]aH'{fu-#MBΒBɲjkV{1A fe .SDRe00I <;TܡG#ټ)SNqm5FnD?Y;$ &/lbӻϩDl=9 AM4RibA-cיִ=|(!E1 . e4* qI CK m}HB(UP=+m6&mP8H*%:OV !I yY7/:JΖ9a7tLȞw1g||}`/=hD,̈)920Oj2k@X$WVA'9>y+4i|6gqXArRΒrhV*dh<:V˜#gdͿZ~l & lȘטu&׫0{(]M?Q#|O}9de-ڤ5 2?ڟhـQRJXBeBbp°- rFꅑbc3"ibdH# EWwS \R xDAQXnm+T5"(.Ne&~189Mfp#b3?:f-<V|0ɣch ,NN1VQHP7/K_ɥuZ$UlCqڥMH>N[KT]4fmң=rc<;祡E9~t6]r@S ͈C:HD7z19֫7@3ͷ?[6u7__31V/vsvP涵koD5k ؓV{qB2;{j-u 9|}z̼%(FJ16C}tQQihi@4dʣВ{ Y 4mѾ?JoIVJ6vciTfC>/jKޏizl^-k_>}lqU{d@L$v4)㣃jWlPt٢,6;A #t A9Ub6du+XN덞_ytrkqvݣ5z1*hm/_^:hȲ͐ݢ5EUa!^cai̓u(,թYQKr2XNrZM(mfvOg?S CXkzp)X;h2 Yy>xsOs棴ˣa9+gXeNrGQ::=d& ?ӏ?Կ~oToR1S&b迼cï.ԍMݭɧsQ$,H2 (36F%b-AbCó&Owy5G4ITd}ɺ }-,p;kB2FK$>ry&ׇz:ee\vkeRoVw|ޓO U%@;Q@xzqy,2bȈ!dS@ NX-"R}P[ysJ@$v*Cy|/SdVPx;@Fe)6AZ:Ŏ(dr" X/yR25D]"{C>&ByPsx C35)j.\δ7H ,SM2"x0|X,֞]0iٺFRFdʖZU%eR3 u*de %l(A鬯sqVLbIA<|C}rnnƍkq[tYYkqa#'GobBg2*:M Mֆm* lӉm'G [풑+M(f4vJ\MPQl} ɕ\1WȕЂ\ ecjBr0)dUU{6e?*>rf$grpg0򱜅:Y۫?Vd;:[,G+2|^WehPVAQp51{V#qedm>?6?{Bd/G%qIU %;+G9xQ^}~ #egoUqx'[8YUC`\W?d95:iƊNZd*R[s?OV??^K=Y^^5&rUۿ|o;ʠҊs7J ~_q7KhU?WެGZj+ںufc'nTֲ|~Na}za|AWW[^;j{*Cqvߠ ÏLLpMw&;*Esl3 #Ψ[Fҿ|n,SYW`gjjRv fUYDUyf/Λt~+r;y?dM^vzJ 0O%iM%|4?]mO_j8jf2 |[& ,sq_[ qo'nL$${\xVC8M<[×;DWѾ6|; H;tK3(:jي/.nl]q޲Z%E{6g{>mðɢt/}7/6Yw{}}|Nޟ0D],nA+hB{xCp{{EE:^(ԛuw6=#7HGr7Ob+D^G$kp,#<pNt|;GZư@ > d׫T(='0>M0 3 #Np1"WBF)r5AB CBr!\rȕо/x\ e+xȕIɻ`NgPT\ w%g\YT RZ' ɕd1&\ T\94R+&LgR2r%/"pJ?w(Wɼ^݇ʕKt23BҾr%8J U)ɕKF7@ z B 8՛+ݱ"5 .WU?@U/ZcQB\rgڻ蝧.9c23nt:3L)PӭߤZ/M-;-Ax| v:CpI`ZTf>P>t| C&$W L:$#Wɕ˕PY^GSdqȕB\ s\MPR{kdJpm2Bir%a)Yo+&#W]*rŴa}(Bjre~<`w%ަ"WL\ Ǯ(W`_^33U6v )V\MP R+*v5Ɍ] -˕P/g\XĄ䊁A\LܕǮ,WoFLǢ7ʅDc{tB=L6Sꝱ=-;jp^ G*{Q;iC,=kn:R/X+15B9^ /ˏ\Q7s)G?wY>?H7ORK^viGQu:␒mǫ%Gvz$6UEZ,d|P\ME<88QQh"JAc7([edz)|*uCmЏ^{)&T]/`L2r%ڥ"WBKvr%f\A{ѡr4xdA?\w%.WBa $$W )Jpm2SVBJ(jreW>\10)\ n:Aw,W+|JA>b\7whaޕPjrä: LؕZ\ ;Bp UCJAR+\ < :=* 0Н~8U?Z;\t#ۀ zrwѻ>ddp!{N?5-hw.nYӊ7-1I̱IxU:$c B*>jc%ǘa&$W @%#W:Zc+a NHґ7Jh+\MQRWc*r%/.WB9w')WdWJ'#WЎ~L(r5A)| |:ȕЎJ(\9`]Br(Ӓw%f>\y4@Br6!q=&]yk9\=2t* Rܕt%33(aU@t| m@HHg/'#Wk\*r%.WBi,W+IHЛ0*ճ\MP0:!\1.dV^Y&(W)yW l 1E\ uc+,\\9ȨJm:cW;hؕPj3P[\ґ+\1څu,WoXE8dyMǛJw/.$ŬD:rҚx_|X]}(_g6EuF[NA-/ksXVXZP{zU[tv]5Kn/.x! Z{._NmoX6: Z$se%`P!YK]&xf&~/ы̓*yĶʺyU\)w^,S%ܿ͛WUѲWHnf9oȖpÂ{v8kfKLaA(-%'W=~?~HC-Ļ*+O﯋ ٳOBSZV;됳笒!d}c(lĹ] us'ZUhFkXĪvDuNUܔQ\ڪUm-\t҅wtsu폩ETs`5NԱ TDs F?r^5"H]6pג tLXeUǐ,FKd  slYp66/;eP%xcKBv,| g,X$"}PK[&;<1X$5>]jfjFT,CD ɞ5)rS`w.@Mu+Xud](tGjOM‚w_KͲoG0L3R/3ޅ*m`|ހ)k|FoxJLv **6(:٠-;u9xC9ݶ+`BY @dr*j]v_@2!Zn kBicmu+yX,čd Hqah VCt9Qwk.4˺p^ LlKfa=gAPѪJ(ѽkj $*J'0'U6.H1?'؎qok۪n(WJFdFfCV6p.F9G!H8(k@oB&uTWPN+$_Xp6L:@@=-V(!ȮځpYV Zq2M![ @C9'X !@YQP6{iFn3%!2|kʃ Atof]`&b1S|bDjppR( 3k'DŦ3إ~.J P`S1 E2GSF ͬ#kyGT),d~G@PSaH"r{RUUuA"݌e_{FXy}W@Bb|M6y-d*- J9&tY\D d`ȷ{6 {xWZ9|p.hv!eP.-Q̨KQE#fD4' G1͋BIJ!pB6}*y S9ZˉkN%W]Zսl| mz`-$L>%@uP< *}vtT2Mt%C RUFB5L:fv|YBgԅs՜ hdUΫ V&deZCP9xOy L}$kIu <o 7>pu~V% 9ՠh%QXwV4Zl@=`%!ud14Mh%A pf%z,1.-zro i ߬vAVm W|ZR0UK mry;Xy==nblj/9M{uhș$K7 n=CnNN44zQ63 wwG(v%[Y:6 k*ε%'ՓFho1& P #>62#baQC^:$5\ ty݈XչzE.WRXSA`*1K[Ƞ rw=`-\߼]̰%6vł*'؛XW6b] j wh +U@ 2PlQLjMnGt0RuN` @ QYj0Q3<4)hc+7 ]H 5k׃*M3| R5:gҼLFAL@Zv!;9y}d vCNj-z5FC-6(#qVp*6m4EO0zʤ輐؜?Д n F[=>\2 'Y¨ 8 N’.Cɵa+f(Эxzp4DfSMhpUfs4DZ%t,6FR\`Xr} \^]9]K˯V׋]:0v>[vz?Å޴m,߬ҫ /hN_JqE[Z_=vy{|CW+KK777/ano[[/ަ3f_/֋5%;c];0)K%F o#%\\7M I |L?LW$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nu@ѡ`?37a@9)> p@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nm(`( Dc& VQ@A}' :$+8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@:$#%8q@lZ#> q@8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@$'8 I Nq@:$л[`t) }]P W7/bچKpYbӃV.\:ҟ>Z"I 7 ]\cF+B/CWSNc7{+LWUvtZ' 9XہGO}^>#Yr*mxt涔'5\|ȫo|sM2 aV/^j}ъ<$~G,=/C-~zgr_\Wnޛ8rsb3eGgfӆ\b n 8mXQ-T>obpVnQ w/M728T_UݑN@1oۛxԤ)"N8uE'ʾgj wFƝGqg@3B#tgi/@tE]nУaLWGHWyƁ 27ϧ+B<ҕk;] W+9UA$*0]!]Yoǡ+Ub3Hh:t"0]!]9G=΅P׋Qy/D8t"]#]@~np` 6B0Žt!Oo!*5 ]mn/LWGHW1 @tEpIZeuס+ĩnesaJ/m\ wj[v;;jx7:엥wC$*wC(t,w`i,S/2B=w+mtv_֗o%wQO6mIX%穫&E4f*!d[} u! 2 tĩI|w3]Wnh|Pt>zRY\.e[(Oo)Txh=糄'^M;1|A(O7٥LztCN%q;}t`S`NB;2s|0;#NmЇeΎН)ܒw7 CWk:x"v8BWZK@tE8tEp0tSXx>]Ji6~?e7QBWF[!^]Jf \Y@hc8tJ%%ҕ39]`;$΍BW6$|(H G]\G+Bݡ |*ge@fa~y;t0]!]E?+'7]8̱+ ?Nj7jj7qOǮvB偩]E  )/<^k[C|Oo LJIܤ]8qвKj~G~^m_ôlWO \BԮOc.m*6^;v甦QӞ0[=}yY/OVo/E֗[_-LWkzޗt?U+ݦ,o~|wʹ{tmU?>ޜ*_jdmT'&jR7vj2RJ}yNIq3/ʇ{W<@̏zZMJw|i\ tXoҺ'|%::{yn~rA;S )ʬLLsAlaKټHUn$V6_\q=?I4ݿGZoVuC߭.naN|\[NghunDf)G ׁ`yO# 5-z^>n6/G~3lZ-_0޵q$;FC@slg6U-HCV߯z/$I=X3=ǯ5dRp 4YfЈ%%pʉMJ)o%9U'$Vv (] Pcxy$$G'*yL'o8KEQY10_mn >jO6 h9-Fp_|pFCa3|폕oBi7Vru\.Rt|PiUTi}:x ʕUJ1]E•U4J+J.v{w#i|e `'&T/Ƿ:5!?avL?1pN`4h-_C+Y鰢RX >n&Nx \tr #FZ Iw lAc@ۉ]#EQ Fw-dp ɐMVO[flkcZ^b w_AFcpQ3va dgqH$m).mm o/ghF*x@8M3T$"" ׿In}{ k۵xam۪l &lᶃB (\ ?*5bYx{T 0 $pfju.-XjC^S{Ui^RicttبxE՟[xZxųF;k^~}w7 o+!|QxQz񆍘~fl{ C'dO-֧b` Vvq (E gkVL! -8]dDJ;c0J4 Ev$e13P uBSo0:ȵBHRxH@κs{ YƗ1g$j︳Z]w{nvFI"7L``vB9qtӂTd'F1J%Ğy;azBitf["{a\DP5f 훗otoBa^(XveBUJH]XJu` w ['.8 O#*L9dI&B٬U6Ap\LA(ij"nhq*1 %l"GPJ]y!@9a栠+/w՝ۡ\tNvr0Swo{ѹ:䍆Q;mQ:RkBB)a5Ohl 1;qғ%A Nk9;έG<Xc$H!c9b2I;`)*(";:ؤSÞq"dN( 嬄Lh\v$LDɂYgݹ#5|f NN*> k)d4‰dFŤre-aC䑩OPdv}3}OcC_;^"P[hkHRr@.)ɂ2ܑ*{JQo'v3C8n& ΙHXdIbe( M\ d"Gx>mw)voiߎkbMRdIR"#54.X"-礜B/q&ٞۺP.fXVZP3y(񎙏/쑜IL: C^`B:0Vfd-Ơ3s$D Ui2:Bc_.5qWDEH(v|5@ğ,` 6;4|D-/XQJΠQE_%R6*f&+! jP:ߓps8(WeM ;,3%q1r.%2*D<](1ac뾧`%Mʿڍe~ゔ3!}tDͳA"3$jJ<ē se8Z ijfzsO@֠VK2G *;tΖHv 1 Xen?ݗk fxclShgֻtAHlKY͗҃NIϽAJen_g њ$>f9gƸ V\L* gn!>?Ou,#IAD&02e93:H BlϏꝽ~Ҳ'€j:9jp|\U3?T/ѸOw0xWMfZfJqWB󶞣T ;R +9 YxYݔhozrh磺J-yܑeWCoHKqjCʾ\Ӣ-^x2bHD~خْ[:Ռ݌;$7ĴIh,4]׼UVcVfxɲ"BVM%+<,R%4_'񋚄ot̏H^eVSXjô]So?^||5ou{r_q]"soXDaRYMypXҢOUC+ 'pܾO$}+?߼z~ݫ޼½{߽?Ѻ1dY;}t~~TD>iobo]nM>O *-r IDd "4'+eɩqHBgw4,O"}+nF,Ƚs.e[&s׉::"{CC;&ʜޒK{iϞ2nLB7nnW=y!!Ρr\mxy!5$-#*arUJW=K2)S-{ڄׯ,gWON O`oO޾*ӡsyQЩo/^Ϛ(:z۴Q=LԨj"-MVMx]ZXHBXV]ajzn:sO!q8ڟ5.0^2b:z}OZjV'iNy3PJ1W~%_`be`>)mdR#˨8מiUDmG,fIZ%f0d Jac R:B!PjFNGN$cB; !"cngyp?㏪rǧ9IiHT,5HGRz5/,S:[[]MƪeULQg6yK]j5]O'=(b&~o_~v9\u+Qx*=S]A.*k3UqU@22v:+]sOK:܊^ $vw,V}^^c{L,c05%a`xTQͫoDܷcg$FE*¿zAHbM)() u۔!OD!eS-1jX ee4z`R&5cxx~68^e 60W[2T~7DVЃLr3nhl\D8K7F? ZlI$ChsĀ]L6`vY!zAW،vu{yQb-/\?[޴tsD͹զ1Lc_RϦ 7eZ0 o+Мuyϫ]7D\[':~v[3,177WMuQA|"e&+dDhK,Ze;P){6{1jIK .X`1%`Mj%RsG (Gy,+ ML^K<> O?>h;k=E(ASi1NcacC@CZMʪ! b <* s~]`DʩR`XH̊M͊HeºS"f'y7ae/@~VTBpq4\jZgQ-.̸(:\pq)f2OpY ]Qu\Ѯ=)Lhmaѱ-x@XF9uޏRS5ykOUe);%iq 3+%TٴvV?9oH0[ɷS@,=U3}3S+cr`E@h%t"(&%IyM=yrE؄ A2 ǨU1 ,<_^0Y/5^#&5ghlf!UQؕ0@Ϗ[uxfZ|cHW?b͑T4k: }3)&&ýfQi8.ȠY CiS4:OcE íPKcTG8a3ࣗJH4@لU(Wq'*Co6qn .G})Kyʖ8 *8Vj R$8^aZLBY]i…nCiNLj9+=4<*fDZ(̶~.~?>b>Iņ7HF/|b m`݇9ʪO_?' d:wGV^='Az4Y_럇,X!f*/lH/PzfHט;fzOT즞W$-8z:J;Fb)`ΎR!uК1g69GJ@Z1>PT`#8H=B9DŽRn&x92J#"(h-J)'RG$A\ln.IvOքziu宝G>uX l|t&*S{W{x_6?D}#Ie7?D=+x@\M(v))\=@@`*UW}+avJR2+)l & \%q Jz.=\%)•z B \%qi%j=\%)1•VL}tJҽ!IZWIJّ;Fr+ X`3C Xd&~ɚ2NjE*Y#[ArIzH>_< Q5=6\bHIjTf4Kkehjt]Wqr{rzZŲ%mvs]7mvpwW> x_|T^>|25)X, H1_ſVrq2x٥=;˱79;X* lyi|[r@aWϯjaI%X*^>yNN70W:>jNDݜStJ?7*ԗ gltNҲ 4?CLQ?bp,Epie"aP"f)'8K1 JmFW]+HcוO:A]{ny!@ZҺIW'+Րft%ѷ+cƮ+0uP7+]+ՊvW؅r~n)`Nv%] J(M:A]ņt%4+]V+?v] NPWQk-Jc;pY5Ah􈄁4-drtlz}4:`v˺|w Zs@ʱeWz>Xy7ٹsR3/Y6ķ}"Xc݄Zǎ3n]GZ8ӴWYVkcV箛9*-s& 'v΅r {1L+9r{!NM9 (ZX[׌Vt%aQ4uEmIWb3nPdWBu%&N:A]Z X7+Jh2IW'+kblJWv&%ZhYu%a bӆtȶ]9bh+PNPW|!]y Jft%̩vufNPW-Jc;.ftڠF] %IW'hܐ݂ˮ] nӺ/GWf˦7v:N)s}c?n1Gz J#ˮ]IW;7}pFފ"tm$g]ھBb0o?a j|-o Zk5}Yt#߷eP5k#L-;m ((؆fgfMb}+3%L(Lg6 JmhFWb+cוPIW+r7=WWlc3\Lv%aAv񤫃芽Bnݐ8+݌ָJ(O:A]YO*Ɔt`cѕඓ]'Bi&]𢍾!]Em3‹^5sJhiJNPWއÞ ~LΩv܊֏ܕPi)*uKUFߌ7VtZsͅQtut`kڙ .53Z;SB⤫/FWeӓcg]1xOiwB9c^) Z۫aaavdOIӻ60~RM'rŵM''5 xX `X+!2ւj7V(ikOqň5+] n AJ(ǶAˤ芢N5+Jft%3BKjZPN:E]YRC07+=(h¤ԕSY̳?]cDk}+-);v] r9 ʙ&&2JpsWBG+4uTtA7V]״sẹ?w-tu]EᴥJc; kEWBkxJ]M̧jџ4ڇњ#m7GOڹzL=zyՅUq?߽se` *L\?w4Kyɚէ%#bP0]2./s}/Aqjŝ ^73; ^~v&aiHW!vٗAY(JϓNPWdag{++͜Ʈ+C.']DWlؐ5+ft%Ό]WB9eW'+o)!> 63c-3F9:']DW{ߐdnJ>Bg +P:Sԕ[nnSpۙ mJ(㤫S!] k]+c'52i!)*2ǦUCA5܊G](tn\ɳ}a  hBFiF6te']!ýGXu0|Ar5=|{Ihe(vq˔nkqRR v빕Ch#=V؞ >I1e;g\]-Vt% qNPWx P3\Vt%֎]WBuT%PhHWJpиa4J(y:|Xҹ+;ι+L֏^WBu{eҕkFW\ZƮ+v ;Fp `ɮ43ZäUٝi`ft%] cוP)*zҕ}ASBkF+)rtlzwvkx!3o^ 0: n! ?psw7zD5.QK/n$ _~叏O 5ƾf{b˳_`|6_!-Jܖ-ݫRUׇaG`=v1G98[v/ZQ$c*!FE)ѲP]|6o]mq?f}Be7ү.+>޽|\yiQ '5ѹv/~ L7eiW'>͌WOXhU];}m_]m~y~F!_;35ÀOk0c=`؞xjnvnzhsAKOٙJct=x^lGޠ赵cjd3v-m8(6))6[|mP+{5򸷂\S3{EWVfgBz33AfgFdkJmFW{ekҕ8v]h:t Y6o rvD3{:.D, LnC~TԽ+k;nɒ)>uV'eF.6Zst>.uC%Ub5PLRQی5֧+&2=w *[[C(ӕ[hh=\ ګPmn{g+3dR:;m2'FVk>i}VD *tAZ ZQ3WmP}/:]H $6Ȳ)զX{ks 9[\Lo+CwȤlm՜R}-BR!gљь1N#c6./2:֜9b;yh`fx?]eCuhQJwE{AјD2T0BߦȟA..W44zT>>'ev6+4Q!(}tB?3U:PnRYR&[BI촇c(`%Bq̮'g:F-*kj׳3%T{oTP.hmS wQ:oZG߷D}0Gm{nF;IJB61c7|Jh>6bjivuVÐ QU=KFCGd+q]HJT7¢?iib)*>QlT\U>EC>KZD0f;EʥV!uAfҽb0:*D{SޗѥtǸaڨ1^v-2ZE,0>aw[<%Ol}1ȢbEF+r {kz#sT7grw('juHU$#{FXB}W]6#.Rֽ@2.T-7Tˤc/Ϧ/Ϯt9pu^21+,W7*1(7fFRSrM} 6C֗wI kA(SLΛ`b ~= kUAjRgfCiBW6#Y}BVQ3(JP_NOCT(Ts,+$ &a5$SҮHV p.7̻B]a2Ȍ1Mlom11%$PDd"Hc;fʺ8BMykyjRЍkC5+$s0+ pI!0Y\O% ATa->!cdkpv2Y AܛKȘ:;2J q9eqa1A(xkͲwDbx i~ :-? A!NÄٕ0{FuY"D5ʽ׬ IIl)w=J՛H!Qqrr#*7)}&eɸ$yWM4z1eZ1?V\I36~& 򝜿\-iW!.UQQΘ18r>1&(*s^eIНDN&qVQnaiy۹*+ɕ^-y1 z4r>DX = xuPpxigw8@E9+EkmȶB_3<8ԭc"4)~&)v"Yhm-^^v^MIW}HV*U2P (yf'Z,J茸pΠ (D td!.n:% EЉJL T "Q8TDYUQõ`QypPBHcF8&gو^` A1*'J s_ 4+ A,g#MGhTf%Ei*5^ZVE*V`;h QH6 4]TtaV-mmm˷Xܽyh4PAI{(i‘_޾6AeF쓊U: I;`ж+vsC<܈V]qnQD{}Ф.WRHw]SAA ,3 R4Č]ڂDe{!kQvףbb W6'Mm2X{iP'7 ܭA~-'y/eèIP(EeQAR܌EEH,aoA,W (]IڈJ5X(Z54)hc]H렻TAlR>Cɗ Uk&dd, 6#TK.dP?uN^~d{*W!*:|›ւ LQ kJ#|;XAAVT@9 Z6i_ 2).$VO34%a#[k|T(Z{*5]QzB%aIm(6⊕4 t ^#0A.* om4fSMhpUti X%YxЬ$7_PdpNSJ,}Ml$_PFU~B:CWW4"坩y0B.8! Ջ_M/nϗVUoQbʻvFA.ʎN"e޴Gq槟v tTJliv+!y2Z%؇^=F7C^֒b`7ru 2LI)utzJg?6[N϶koi1ƣ#^./s94ߨp~N.v{.'G﬈݇V{Mƭ-Got9:[V=II=EQc#~BsCraψn8N 0qmtG%NStY۞@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; NutWN уqY8z'~ǽNryIN v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b':B V N Dl 1@; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@tsHN   Ew :A':@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; tJN?ig[eҬWm~P'0XXGe~5[>1.b\(l\:'Kod0j3"F NWұZ]`gCWw8pBձPzN .8f(tʭ|LW'HW^G Mw";]Jtut+Z C+BPbe:AkDW+p~bQtEh_]J箾\z&J=9]'tDw`eZΤ躘j)ckQGiQ}>Slf[&Mߙ"bm](ʾG1Q*?`C7r5.f;bmS:DamIP:#1:\)P3Bb_JN:SA;DW\BWnôJ/Ntp`?v8 S߆i?;]ʄU á++P1H( S+1|^+kmNWr[N\_$KW ]\iBW]cd*tcnH 'ƃzy Ԛ* ՐACW (bۡxϥgBXk^xsWUO{"W~(%AWK۹(||M`Lƨ"mϛ Z=Wse 6i vN Fcơh Bk B#kRyvnǼZ]y@WE%~'t*ͩvh}ߎqիWXeȉ=^=r5_臐FކΣzҺ"MPz'Cz䛑ѹ~VRlFaz|v5>~$ȟO?~xyնQVly9q-QXa4P@:\Gf*=a1Pn_a5PMs/\N4+PDE\TR wI^Tб \.R%9y*JYEj{{>/[s&:$=뤬 }FꍑTHمuJyLgkIkiIٙáuYkvb8tp/kvzF\G vqI^fP7lgM_ݰzCLq>ްLz _lѤ tS/w#fU !O?#$exzCMI^m=mہ~ uWge;agEr6#qQXj.9ZrO3p{22ܶ IR~RBQ$bAꂶ%+Ia'K͵blͳ$^!&Y9ިB> JQF>vpi>p =lW|i w*tPjmjߜ 83{)k>)@ @zeV)px{LoO?qǘT%UR^6,s{JkWWh%5*((eu0USN&b523™TL8W.lƞ&7g</ȢHZK9ŦUt6FDLxFkئ.$//&^*]Uv3"vu 8;X<4䁓OnQ+B]eK}T%gul%ktdtGg6l/I|c]_?8X<4;ԁCIЫm%Md%%%[%}rs9&Z?;٫Wy˥>zm |I3ɰrY0 I0Ku{h^eѦM`J8z>mvjF/ݤr|( }@\1D^/ dTLӷ˜cQKj h9]]N;+xj7af6m9zp5rnzOq|Yۉ)c%i_yD4ko6ń]ǿm&ǮwƗ˽]A{Wm (OʃDJ:k[{eTh2P $mOIOMGԁ6AYԚ#i&/%HCZVzfxD$ʞ@əꞪꧪ A2 ǨU1kӀ | \KȬGBQiSlGt\]y- 51L:w[@fT)AB8l $Eă18HAq%u4PhhF54")9\*"Q* J3`DTAy*{A]d(`ܳ UHq:`4w.m*C>-Ю]S Tϧ+F5ւJCO)YǚnpWAGF Q ֆ XVFr#6ݜ/FiBK(l6 N* ^cC >?Pq5BiDp׃QT'/4;+>_>*/0̋4"\8a>[)i:P5^sҳwyڀ.7$ӥ+ַ'$+-2yhR?s5T#45q+G $F,UbЧY.Rq#Q꿩;#=.?Ok=6~nIG)05s}~);s]vc43W"vɵV,v[! ``ΤSw`"\A1y tu^ߛ F-Ni& -,IV+9a:$7*yؿ|01,Q&9h#ӎEb[9kꀐ4Ӯ ßca枡Ւ-<TΪɃ ?KXy[x9 behJ̥n[gc;zޯQY BZX[\ YLf~9 QL`^Oٹ6uښ`emnuuNRևWҗ)eIL\JJWxS^1Ь"0Ld05Ł=q7fI-0*e*-RWVKI˞)\._OֵJw~ZE*0: M$!)O r95TTI$^u;s9?A 9~M?>=})&駷/(8@kHLޛ"@{p_]rݗ.7FSܝ#߹}(,`E!Zd(h0iT8KqK}Ds HNh9R%dI >[P]U#2\*밊%%V5S*QVo"XLN~Ǥrg^ƼPQ {CBiM(0 KG=pO@AVHƆZ!lû埻oښzW;Zw%T4<\see>rFo bR9cpenL )&뽁(M7\k0ax=xg-h=Zt1y[s@ܥ{ݳU#~?'"j/!skt>4A+N9.xT:N1U(}l+51F: lp6HF#Rz\TknXe 4Fvd>aOL#) w[6LKEK8;})oOm èX9]M]nݪ^<.l䍭ɡT(|.c3q-9O{œ&@miqȝ(gD5rIIiAC졵#Ly=vкs4D~f {w4$B TzĐjp $fi$)Cw"$X`T9F$㑅H1ꥦI:oSh:-H7#7 q`z]NJGc7;*;)y?E';q8[frz_Yxo@H*ϼ"Fcb|:26@S\SAYXk޲xfV1 `6 !)lNϩ vhA*:BxR]`yKx&PO%Q%B w y8EʕlFsRЗº-mG!mD 5muv`;^jB' ݬnE]7ci jIJXKQ[/Fb2I;Xw9A-dG_z)gs^͟w^Qq 'd_K0e RMoN.jħRTϏ0!κpa<4!ޔ,2;tGYצyoܣّӭtˆkGNkZ&~C"m];ZE^{]]U}PE:{2wՓތw;|^W7]r0~MJ.lˎS‡L3'E&so\jFtwM(e0so?9~|N$ V 픬YP*RVoObXˀ Re~ t\I EdR2"C!{B{jgDD͸;754M]TEF3^iLx#259+ˎ{i/c$6+!3&YFyy/R 7y(QKl8L"y]+Ήu|Aמ~3n{l +7‭iU"bK(F[ h_4 ^0m$zyLO]&Pm?Nۛmz`^H'orA?6 qX<#xoAtX ZLm4g\#8מiU+wm ewԏa퍃A5~J(R&8"Vϐ)QCǀ-3:]nW%lٍFlKF42yᔠ!W鑁P+KĘKs' A#& 3#0s0a(eY ZrRS;sG=жe( 'zgkzpBYST*FHQ29z1Wj$UZr}NVlIm$b;f3Z@r7%D0"!)PDcSw WI";v"B31RHprZ \D"FDп)Y=Ltκ'd̄mdI:HG~XLY@,&'b5/M^gS i; *|d *.7B+yf6%Ʃ(ku.t[Bh )(m,Bt:QY %:hW%-6yTo2쎱ȑQbg"4$ܕS"BtFDK4K'2A%Xa,C.VfZ.D.ԍk#EN`r Ap 3K': k˫`%R*jrտY)FYɬi{|E PE,Yi^had07CvfyuN1X3 7{}/Tԫmv:Ptep4غ?v?b׃$q)8?i҂ڮ=ܧI)brBY2I8J{)\8UvVMRsLbO[_k{K]A}D֢nx|Y7`&W2~gt= k=7?}_׃Df 9vtX67C\ nG">dq=G$”gB5!"hai)<1Δ <0n#8Y;[o+ jHIʽMeiylV()5khm:%* *{xz|"f4_ht{<bg5_c}ٷ؇E1m w9ʢW_.z/) 8q SuIϐ{av sכ)d4%S,wf 5f3别͇ް.+Z8H'`?.W& G{KʢZx T'Z?-.BQo\AS>Iw2lֆydR#-8%6HvH©I@e$vFksN_"(W zC诠PɒCe^]yXsm3TB-4|. `'VZS8{Hq_>_^XF4̩8J9JQfilu?egf&SN )'H^Ȁ@=~otӶ$z-=zo[/Oף l%|bAPG PGaCIELqA6:W6jPvH DFxI.Rg:EKbLx-Žm؂g?g}v)*FW򅘥D*S(?_ 0G(/lxNV*isUw9:t@3%#:QTBP#6%țhdU6boi&r:g(@ `T<rkʱgM8.Dhps,G*sf)[[[Pr 8>ĸ:weBUfv슦dzMFU[8+9#Rц8*ݫZ}ק-?zrrn f^8`^9.{-ag\-noruٷ!W<2o#믮hx]nW6K&9yɭ&g.>mpݓPppqrg8p6,sk qs6Ԕv -S Pve(} ^45S=ZZRd!;A%DzJ7S<)#<*awFZ QrAEN7M@.1yJ?.?]$Dx',Wx4HioZPNu1-#q@D7j2D HG4 2O}F\k)EoPgƳ_-NP1=٨ɼWR$v`ٰe/@UFQyu BqA)%&dpNyjvKR܌H-5[{*s$KC!DCV!ǂAV R6!H5U4cS, c!kpX83Ξk.@o~7v\Fl1׌%O $!hQRE@R &f)s0dg#9YM.$^fJ6G&10L>&n:)qV# nJ;6EmhEփ] #4ޚ,W Gws,"J`g^htF0 "]dR@yT5!/YdKtcH$HMk7&jfI:):zUl\^],3/ǐeE{δu" ZR\kbTbZINXHNGhTR{ժ8c3{=iwIiv׽CcxZ50 +ߎqG?h|r|7K~Z~=~kvΊM*{fq^O>8sqѬx9wu"Gv27R|+]b{{xEϔrhx73kO-Iƛq43l̅DJǭOǚs 5F86PJݳeVWwTd۬/UQwU{ۿh~=DME#zx]?(xg*], o;Co&U4yb*{Md#)+Le_~J%*Y?bW_.|/§޼}.鿾wN)ӧoѝ Qq.zV0~r+UT"jmDTJ ]E/N[h!ByLBUj)et|>aӕR5B!AStgl@֞Ug=Ħ;MnmfuһO.3]o[^h)`IU18j)RL Ô Yqhϝ!E!`ѣ>hvC'Uw>lf:'~}-֝28V05{ͼقo;r[,‹J*0N KV1 d#A|j)PFs+>Ѻ Т ЕTQR!IuI "Ȃ轳&+r@3 ̀gi&1RQIB1 $8aY6K}4ZmSCGK?/ n.koH{6}VaTeQWJ9`jGhl4![HpqV'$)$6ԁO;9;Q)$M>'O1FrHQGeB@f>J/xZhb3]$Îqbĭ44,M8&t@0Z )F! 杨,gȹ\xX=JRBbN,D͜2D3Jc8Gƃ& @<>4 6t"0x&~%ѣL>9{h) URF)Nk AȨbZՠV:;XAUʹ)&oΞE$(R\n rc Oq~Һ!M+,)*]UkgR|7XGws1Di4-d +eZq72.GiM@^L3mEn=iUcCWU#ZyyּV¥RxRl:@ M*3K r0(+0ǶHG59T:t3OBB4 iC>3 QO ]8^*=~Y*M?ƍ ջ a(h,4hX߁td ?8^d(!ot=3{Z~X!2 sR(FZ6p0xM(Amp>!JuYreR,pژ!ChD[ܦj{rŌ?;ẅ́8SpBݑ+V(?D/y"C:gT+x.5 ^ϕ6讫sea6eNlIGa9w&fR5;w6c$ChU-@_ޥgsռqeA}s]Ўudj$ ug"O"34Լ.1Ǵ.3s@Q*˔t4x [JzlW#Mdžgg`/cSFr h'ko\Ojѻd-$Ep,҅Ci 9&`*') eRܹ*4R#^n[BG$ŀ2%cpC啾;z[.js,t절6bȳU<$)Fc`&,i\jNp"ι`dd3+*o)~g4NT=4s$Z8FRbH'&1$|С2>$8 ;SEۍƝ 8NXs<{W1I&1J'BFenR#K^+P3q<,5ި'-" uhCf?jn#^eHR %^x1h{dO3W9Ĝ!$Ϛqe (Bv%R 4Bry Ivjjjz7;ȗ wO"w'|,h~~_HWl_2L̿?ӂT<(1O׎s<8_;$ .&4ـYFE0 Bٰ^"]W<,q?~?,M_#x4`u|u@<'9S;yAh?͹\]w7G+zW_~?)NڮWߗɊraŔLiͿrȒu̙e]].?. 慤n-S#FWsϿj|]daW9Xu.R|!b`4 08ELnޕJd ^G=5 ggK-_ b54-SMay^^ z`|̂c/eџt>Ѡ?}rGD~4bf!}8񢼚./#fQ-.s:OB/wqVa<; 4$ G: y=xE ymWa-׼>fdqBl ,>2Iq/th1v^r1w2.Q\qύ. g7u9$ 5<0Pz׏y#HͷU1h)9!=cR =@H>!زy8'-l_dS ! _iyH$I0eg޸BhfxPND ҄&Gp:Jp`ovUܺ*[VƅUYbP*E(nOTBQR7Hh.Jdga߭<<~\\^e?r1 `ym뉉M퇱ct6!&'Ht ޮI9;s^kU v&0M&!??M>=t5lXu7ylh(2T%֣/k]=*ޢܼRsg22:xa1Y%(M dcDp )F4*: "!6Zm*mrd @_{Sj+{髑s^n gK@6TkoKbI$uX)hmBO|lq uKf_ALWJ"9ajth26Ҫ @[HCR ǤjUʡts ߊªj {-,@U0c)4q^)唄1bE:1#MT\bWVQ t-_ճ.814Ć8E>8mА(B'fGeZ{8_!^lƪ~ 0vdܽFp|0)іI؋[=$%J֐4ՔIk yz@Ē9}X35Kd\~bܫ3vQo6awJQ]h:`) QiD*xeVei0jxK 24wo,{BY>)euRi*\vu5A#ZYk%!ts ʵ"{ڔ[OM@[o~ 쑐Zn%=ŬH1%lF1%=*Xn/X+#3bnR}>ݨFO ^7(F ֌;wgO?~J# effr2^@䩬FN9D8J1JY͑o|s1l,3k0E !6l޹j%D[Οng4-J e8=d#Dga2 4hK^hDd+ZpLfoS$sI3' ֜ˠ#"q-8-A d<{\Oz+޳J%rg4ħ"fs}cOD%r|VI{0M>.Y56jL$ϐ )DߖKs=2A8-CL+LH<^J{*3jlsTweƖ#tp*@:-~8 ㋏?_t:"jCft0 $,fR\vr1]\fvg-<[>8hP+kh+f?,LJ O%fp/E/J) l^]缢夋닄JM(Dx7\S/zV-y2x`hLD#Y@Qs>%0sH!HOL-1`gd6/Y}+,$d$ )AFx`BeC&zCMmf6޺>g'ޟae/_<UQy}|R r4'25{)R!27Oʨ S8net1{Oh0W9'IԢ1[ϣKN1.HFjGz\VӌmPTBc{ެ^^fQdbvpiP?<w~>p>|O>h&C"k,f6ܙ2YkRL3䔢^8csuSjHP=96AWruv&*!hS;8ۏGqԮۢ6VFmQg4> F"18b kA05Ȋt:i7uuUuVӒmqQWEb+5^2 M*@DZ-mq)n{2 q&9{.pqovl2pV1 0UAp}׊~TU1o`CHROby;Oj~%`8}O7*P MǓ-%oJWcw?(Izp|Iz.RW'Z*>..Z|@XYKs?md,y }Ц3ZHcYԻ&f:?Ate hjav\'Wګ_$f0O-.`G3k.&"1m4|,j2J:|YBԋpyHџۿzTx`v&'].?6ZDg 8}cX%\ tGk.YdKm%NcݟXd΄N:xf'^#Ks%XtɔzeG dN7(4\j\Js5  ol@Oa2غD}Ԗh!׫ax M( lUoi, CCK[Cjr7˚|+Vd(C7dAKm~OѢ B;dH<#EIF#rsÝ'jLʋX9jcK>ɌDh1A}6Ҿfqzc'c,힓DdR(kw&0.;[IOT4L% D7]˄ fv&3FF \0{Ӽ'%,.>UMC.xP5wlj)V;6Ps~UŚڪ|5܊~+#(ZU_[2<H19i9ai`xm,{h%r[dfbmSsb6uiEk0+ji2D'RTYc"ΙI"dn9v2.E5!pz27 آOqY5d愘Zb0{5 r}tIͯwQ(Ը/G*6'zKYc}OgnU1ay ˋ6cw[jCrsN {Q%8Snwovu`xWzks$2R܍4mld:e&qD &٩d4F6lv8a *y{@p7}u9Be껉[{[ٶrBƻ(7{~R(=*FlP ؘDÅ 2G(J%5S![P"d=BH)y! #dy{HpKb{0pUW$"%g=\=C+Xss0pEHZ pUTgW(u@pErWEZ%Zp JJf4pUVp0pUîvUi 3JIWWJJ`ઈWEZ{zpB\N",U"-}"=\=G­9OkmZ<6}p&Ӣt͜87}l`t  &78h%pi^НZ,:Qx|16>sDqr.2ҕRhuĆIЛH[ZOC1`fgZ xr<,!ƝIՆ#yV!9aƈ;ȥ Y['g U6>jœMxANyBs B:0hR&/Xgn)AX"d)ڌLe Q5Ƅ̽k ux;Yy~떅!:i7S̋v]?>; )jOkS`$/}%ȫdթz,zU鱄&D Q6Y ד]ܒE͕%"q7`VD'26e`X6F@Uˑ{Bar2cʂ[\eVzNΈL%sle[M KM dvR8>·]:";ੇB?Q<]9o"FÝeEB2R(U9$hAs$6@=".ո0DE KIjol-c 68nE&#eX$C Lc*Ĉ:e,|42\q dwi$ ʪF>F=-K>)1pT^ݮTOf0>]}bf+jtMmaOǸ ]{_FW~ tq~Kv/ׯ_'s5KYe9h)"ޛś)3yfS:__8Oq[z7<ꊻKԿSR_bs%.1x M=Aɔ F9x`s˿)}T2ׯ(V[Ŭ'zpm3p!Hr0y}}ߚﻅMTn!1(ϛ<ɳ؜ӊs繗ü1A@ـMh .\`CS?vŏC|qe~4."lz$i^Fk^q߾Z T(||pI6 #Z9.jRIV*XƜs.h% Ȉp*yep2F/kOI־h]L7vj*vۻƅo ~j`yxr<$`A|X ZL^}tx*PP*vmb`R`Wٕ,6gWW0,*ў,îJD١ٝFlO@v'YYƒVV:/4*=2JseS;aNro s/e$ BWc/59Rae31gO$!Re1`*(&z&GLb|qHNER*A]M, _QHѩS9e_e3o3qWvmӅٳ-Ѱ??JlhMв1 $ WkCQI>=SP)*2U}pػRodeQݣ(As?ɋS]^aqIBe47RIu.>-()u_-r/7wvXZa@L{LW+va O1NA?' `Ԑd<?.-JV _d"WScgӂ y^N~vzop}(-G}v^н1g8__@m};^rIl𚨂+MEJ[e5R5߹ɺ&!ZqfEȍ5ZtՅ^y2L 䳫UW(ghުul/ixo]fW^8s>G כ^f{~ΣAޮKӳ ӛ=&[卛UsvS7_|ݩ-˴?iY ~ncy6 )ivP/'rRAp)CH+"VOp_9a8$DzJIOϨ F䝑H\AJ'H{n3e_c)|G&_;~8:* 0o4HioZPNu1-#q@D2dٳHkA+U橏p((ڋ">Qc]jЯY1=9jg݆fw@UgQ=S2&Ǽ4;%q)}tymFNBP ObPʦ"JR9];04=4*$X1$R*&4R-c1q[rX-,B^#;xMisOߜP`t5?}+ʣkO $!8[ '#8'qMJ'adFsl6y6'2TD02W.&vFð@1jҎ]6YnxH oM]XjyNuR%+ªFp慦N'` UEaR!DȐR@yThkB$^ȴcBx$QY[g=I D6bǮQEdH[S܂ xq H&y҂CRYb*J$"t&2ZحHȍ#B jT9nJ3.8!ʌIx#fr+  .v[J4vh'ȥCuӒ]*leg;QE ?|Gja2M:kALRrcKiǮPz7{V0(0MAp]3E?RMTnm/&o|@?61"ẤpuUzlN58 I7Ԛ؝ 8"D0"!)P.!$F]_R[I";et ̈́RH#vB;E^޹ MPskz0sS]rYpq.~]Ґ )?a< w0Y ϕTQpU< qL3kFq)4Z't[Bh )(mlnGt:QYXZ{UpI-&C'Hw& P- @ud2}.DDΈhf DMe w'ɲɂT5]jrw+NÎbQ6N"LyqJH@(Y"*;(;WRHKyNh)ցtί_o 3cs\ ;f'd4 mmVZ8A}8KtbHr|) <1TÜKSyz_v[v)ٷ=-C%(aqJJMd3B9x`bnAru\݅kS/GP[y{Q^B4*ay5!"\>ꢿ7pt5\9/6[Vob-ׯIO0!ߺ1jx=9zVlTe [Yc3A=A(!š^NN=0UyüL8:S]APqj5PpWYnfޞvw=($=w4BBǝL&kmG8zDs,zւS"us!(\Xύ$!JʘT&H@O{ZĜS?SRt_)qBWKːvU::pcWఀQl0Z76 3ALɁ#~heq?GYZmQWf8m= s۪Wφ%v= M(s2JrδRYi:ܱ-([DO `2EJP"Qȓk0g x ݑ)A2qS+ۮojVm:k[+PisBo$gEM c0 < $ 9B&N'+lt\'x?MeN'2V Em+'AQq:bpvC+ |Tܺ;+Ϟ*E2%ʻ1WT6-Y 8=|NQ^p@$V̧Z5 ]TwItOZ#U,K{JD*φ^mjO\֨n9jmqOqmH[E[=:2%}?!FpզJ ?{WƑ@8`7]-98$ î L"DY(%rdqq zK"o *{Y)|m8Nկ=S9T:- 7%gacrH(ک& kaݬ*nU&l.,DEL#)MZ(^qEO!'7Zm#z^KeQcY,&qEn> rR"(cv\0 XA(֎I*XE0Jd 򶢼3 YY^SwOd4dVoYܛm<PкzHySK8ljh-VB|5($khG_c%>sm}15de(c"`Bށ-Z%$(H&D^$>m4fl4`R0 ơ5TP@T\t809Rv|/{mOsY.$͞*khSc4Q y2 u 'HRQ3n3T]+P(șATlc(ЌV `B;ٲz_5:>]#R sΨU)IYyza})2G-g NW6Uj_m!3O?;-'}vg|hk9KMOrI&G5B=(;Stojfjŋa'~RGj]1LHEx s&nP eyUzn'Vc/.vUk2k孶0?mgN;mTD%)05E\ l"q3a^D]皖[EOvEZM|JP&2 2KyZ{hx}:ua$J0lQ=Xg׻IȦ!ZXbO-얛ll ?'$&Ol2o[9QNT>?cK ($UpuRd : ({nv'e#,"$+(P̠5t/ĈR(>K|dLX'hi{o.X]y-p\ >wJd-ܣNJ&z@TgKGrt>s8cKqx9AS,8k.IQXZnA/ʫN)d#5iU; Iifx8Gx!N]5  P/K>9R^F'ףK皚pM(qr f2@ cՆ4S&u9Ԯ"mS81s+RL(]%Xg69Ι\f{(arc" i)(%P[ɊͼC*v@V|z/ }fFsw%K&TLs@^ 3yٵ`X D8y9̜ Ff!kqx&9Ǥyt7G~SNoĽ I|nh^ۋOzxQNHqLg Ay%NpFᾱVTէv`⏍Vs48HEh5x9f3h^vD:24|@\wp"gS^WБ쵄,"Dkro0[e$%sZY٭׶d IydU9w:M:+ >eiRv|ӂ42PG9|E`g1?_M3r{?'oeG?vq,)y;K)Q??}vhw~V(gn4,_^LAbtV#jg$x "_i׳t9?)MF 7LVi)8NGxoA^ X8^9}-džflwӯGͷ_4Sw5}O˓hEͯVgeh2%Clzwȳ4"M;nT pt_4V-A &ؿJcdބ*|Ra NŗZ)kgsB9ns[_{CzukID=̥2(E2gsɩ#}brp/m,"YLt>o.(c⬤30PF^UN]ҽ-8QѠpw4=:IbvBt\9JCE?w:|hˊuPI>^~&H);GRqJ01HI6TOڡ%g\tZ&k 3Miowvzh nd4313]B"tL5&I+'J·DNĭw1n(7֫]8|: Đ(c1(Ym ٤6Y7:C:v ]_Xam--VՋ'Zpq.^ Jmfz tXZoEռ?BULjyd U||15!q γ蚋YfR;eԂ/}{~s&~_^ &]M( ~De <}0ǣA5U>_/n':f}.)|RUF|]MYPvyI,N{m_ŞKVB]D5톪Tu}}\iZrD Yg+K8/9J3\rV*:DWx "3tEhuBzT{z>t%1K3tEpBWrtE(ҕN]!`ͻJpEgO NWR!]iRv3tEę ]5ͮT}9ҕ#Y+lDgZ":]!JyOWϐ,8d ]+Bخ4=]=K'iuw+ڝ;M(/zz>tEśyp+k:"tQ_]{zϜ2KdgIͶ$v?TiaS'釡}u4 U)@ӢM_=gVj+ >9D|3~?Nd0ʺj)ewv>Z0|IQ(eS#j|qvT_?Sz@6!DԐv)ZHpL;[t[ 3&d<7ӗRJO{sJn68tv#%+ОיR\tpuZD_Ŗ)lcHjh{%̜M*aL@jKEr!P\HX:-&cjsrCvuƷ"xCkoY|+4.0ݡ+XWjtE(ҕb]`:CW \":]!JG%]i>c:DWXuǺ"3tEhu" {zte:DWBg ]!Z+B)zzteּKtE ]\BWzgYHX:Ix8n جhu'$xG5޽)IENg_꛾QWI'} <1V+HE:sbb ߻9z@cs[9h(aaI^Y矾W3M\W>vG>\O$rwn2U7 )l&´REk.5+G-y"EJ^,!J3^9{3͊Ž`@7 nmnZUY&mulK56IC[ۺgF|sIEʲrQLeDK ;i$ƊMBZ%瑩1ȐCZ$SDlMYxӖ iHUFő?G76ɭ>}4Me}HlMs$G4&8Mq}{OP>Ɵ j g3A }Ԟ}m|:Xj5:7ӤhqS]u9jTt&?Gl<@chC]1bs}piX#y;4GĪ?,8ԮIUuizɍg7ol00XkVbNg[LfnʏlU #C Y7bwM%XھM7ߚMQrPk)gomb]vYk)R]=p-Sw8Q[i ڟW:u#ʺ:е3]miiødHG9z~?¬$];f\0:JjSпkVL! -8]dDJ;c0J4 E$e13P J)"dF&r9(h95rnD 8/"a~;FuT-wAnKdơ#BXsNZE\) Z#f(Wtḹ6y`!/L/RMkJ jkҫivw !r殛Cy,;mY3!RLzk3S&cv:z/'K;@Bb^d'I"y'jI1d Rr>GL&i:E2AElS6RG4z*#ѵd3N)!¡I (Ye9k[Y#W~b8} ǐۈZJp"V*aŤre-aC䑩OPɠAu"Ц'W~b>X.RHm QrX3^sZ @J }HFj ji:;i)3G9X<)?ܸ.~fȍyfwoЛwq0ap@"K Ct,Gn')OZ7ɩ,rhEqLpz1z*-ZQxJH$]gx>jw)^߃ 64z=qLIYC 1b)h\,sRf!8l'myD1e>d_5񖅏υ⑜IL: C^`B:0VG-ɦ3ZhAgH#9 [%pt/W:D!F'Ť,)co#jR7MyUimnpX_עoyҝ) ('.䉂Mo| I7'<3B4gU@uMJfZqWBY?<{h +HH82]f IGU&ɵW5{bsѬ_k=6ehx1^0s+UrCNpZ ?49K1IV,tV0ыVtg5~ò~d#$nG%-azMuJ 죷9z%S6y&%[:'{0b'He: dJ}=A#/)ƀ̛mbvrV+r[/9*wwz 7$ Bgu=:iIa[<{܁-.6/h9>U{Vg]ʜ歫Y_uvu%gbANNgVջ]AUiY^-_nWȝ#)}]È8sufyCB 86 X=Y;M7;GQ5bo4Ρ͑.+*d9nUҸ3":+I /2&~8N )\tH_nT^UcҒ[)U?_|&C\k5Ud (BBK,^,soRy/ 3f${*F$#%|w?*?:8bҡH.Y 9l:-܅n? ym'|9.:[|$!;$‰Oy;8i!\+C-( , i% fa{EuA<6@$Sȥ2F,7ʋ9"xU*!{`YsjSbCP%f$Qy dd9ySF xˠ5rn} j4垆)1) X/y'9'n2+M] h˲ @Bg 1$egg옚%ɀM2=!9qD%l90YK`r,hb.T|RQD'lrPZm!{nCܤ^~Ǥr!<_q)Bd:A`y/$1 V$9FFJa'c{$ccv dӻ?wnwu'r iTzlgSzuOB^z EWzk,=Oɪw-ao4-::Ui ;`uPɛU`!e F廚]٥]Wz]̬]@ '`R 5(As !31(cvH4l>Q%ϊQJ';R15vٟȹB{c-&)];zB9P1wWzߧy?gWg6EԞRsL8:kHnQժF =edaruJ=K2tSW@UnB B4v @(Awg~BH>.Su`xr}7B;(|ǩ*)lԨi %H:GK&s2=k-b3OOS=;UOy4ü4JVD{D˔!E@'JbEmC9 :ӊ)p=~NmSO9>{ SFPQd3eH#׎o ǜ<,hO0{֫b ;-^@jXl /ċm@73~jCfye2X&R9":غ}9ϑ'-ytYC'NF[Wِ DdmpI#GƿY"Jm)ڞI2:j%Dp_' QE'G~7l[<L7~MafOǴ /Jk?K '^h,ɢ5;lXSA{{8*xH?~9-ޞUwg/t-zxuȫѿgޜ<Ozܘ>U۞lݧM>~4ihzūr``d\XvzYy1%$S9E<37Ӝ,[]:_]rv =Pӯ}xJb$>oYZNiiXzAi5SMM#2#wؤ۪X c, ٖF,q8n^:(ZrO3HXyϡ,Ӱ}mHaqzG#V³8k3'gkUoB!%/-.MV'#pч&1TU?P˄&,q[Db5D5mFEO/U|`pKwl c[HxC[/lKen`b[.ŇUS'kB%6&Jĩ4r-q [һ{ܫKg6$O߽c?ť χE\M,Sgrf۟N '[zZ:WT{ݸgN^Aܖ=qwĂϓI^ /x ,9Eэ>zkfs4|:ڼLoж&)FS$acp_ qZzd,x!$DRGMQ*IDzϗ5Ti'CNsi/3iш,$$9hקJ!j:F_I7v"_=|wC+ۅǪ2]-=3}Q*'OK<--yZZg Aꧥ[o7Gרk[4d¨\ƁsL"f~5tPwWba rٹu$`څ$-$Y.R*"]Ir6;a#7&t}l&Y>#t)dlV1I-.wɵߑuQ5)tyOK tS SǓq;Lϊ[YW h]1bz{u*g 8{)vپxsq憩`V 4ĵ ,7.'/!lm+dr6]3dw=;Rߐ"5XpTdֆrJRz&yY 2(C.w!OX].s`RØ8ډet*5Me2"OѼfQ*1sk9~å֑53ҵTO7 Ruauao}]h5D$0J5D&.yN Ii[5DjȾC/5X=v~uF ?%:a5|!6>hTcR$!ehu^>%#kړyJ;1>~-``հ܆ÅOȢvK˾ݹSH.Jscn ocuV7VC5KP\bJɘJ#-je" *z5a$9pu? lLizRb_.Gw1@Lt$h@tevReyX<dI=TdVGKASY*儊[a&7QQ:(&,6Nd,%^n|!Mѓ>$Jf1Z̑t^ۈĩ@t Y%}g"Qin#Q9"'Xci3q\003_1ğ Z+l)43'蔀ᤍ^MGq/P%b8>NgTF&O.>]LiJX ֶ2Ynٛz>"糤xdd$:u,pkob ҩɄXl0-V'`~nM_PJvz4{ň ֫7落tS5Is1NRB;M?PncŜ/|_ Sy'ţFk/m0ml5լ>~J"k"z㛋:jƏ7M?)w!'s ^Ur¾2eiPft/[J , 둏^kt+!76e?-555\<Bs^rw3"+r=>}M{ɷ7`?y06|6d"جcS&XkA;3(K!H-V2Su&` fJbHߑ"$B48EZ3*Ojl%xofwN]`;'vV=lX/lAeEUEGED-slҚQFǜJL> SFe!e"Q(2:ǘh0 I蘭ѥhrST\I9R8ۑR iƮX(+cp˫b9u:f.&;zCFg'/O҈29 cMF8CffkR\s=rJAJƼt,"42uCq\Lz/()K xEQgW#V2 TFjlGl?%3]K;vEmУvonY4ٖX+\ d,JД ,#_WWÌJ/H"L!L&&HEc*~#R1cdꤝَS?g(D""VFDGčy^L~+z-hm(A٠ANmY褼V4?$x][S7Ndjp.ydID&yHn!1nk '\3묦%⢮7j<޵LPN?M*fmqm=el&9{>p3xXK;vCSnx}4MIp}+e?Z0)կ 2Foډ2aR{۶NɠHy{ K?09?wrzkN 7?{+,Lo?VY Ѻ ͔iL74 |;(>)NvN5)R$E$ Jc&n:&!x/Il!(Ƅ^K(H4.%N%hReKIuGmyچ5q@d.pe'%"Ihhqaݡz:KG?3egɤaAjC*xO*,!J0¹Y@ -V;g"(/c6YiȠPdN"31hLϦv"s5q:An&7O.MRH.!v&c%`3[LIOT4\%Je,,䢗Ǜ9y¼v?0+i14c]##d5*u%,3]<}RW+ M{btNk~k)N_BV3xt.RLHy[j䬍@0qG#pK9a|?4V]n0L׃0 ͇(OCpK255Q,?].GYI.Cb{LNO..YpeiOZް'?||Wx"daM0~b>Gy zTyp2:K62G͔Ơ6TANPN4Cn 0x:v:[0˾}ؕwbB55RKn=%Id]P+e28˓6BL4sYFqE5Z槳W9&~Ȉ]l}U-bZ6*&L\aKWkZMbj귩[޵X tl~G5s5i벋x_^g13o k,כu5BqL˔X{?nOdҖ)a33] VwdঁqEX01z7EV8v=C$H!xOJYDҫL윋I$@֜(%ga ȅ肴.E \e}#:pϭ1&dὦB+Huk=Lݬ t(C%tҔKԆ|YI+qKxzjy7>zLL?E(c\rWC\trR@_u V]撣]8n7ta\ |FtXc#YƆ kylhͅS{o;-2 r2u&?YI`(3D$0-iƯ,N:ʇ+qt1>:~63j=wuy%>S$CRz(RH 1]U:JdfVW-{q||"E. ,|tH1 G1-"|_2p[݂23 (]N29`%2de[{)pJsT9 Mb76] ݥTԼTԸlM>Fw(x%τ`:e"`ȸ2nC([)QFpx˹qn5ɷ-u%xr\L *H#uE.gF IqcAl#Ⱥ 0e EBjP%d r.F hm W!0тdP:(8ƆsC>o6+oJ5T91bEURZS]Ԏi'"DCCoQ,$ {LP@N?1[4eEZeG)l#A)\nJ-j jiՓZ=T2TQɡH1 I\8:Rm,2O6~!4 F$oG#cѮj&v= &4Xi-RLAh1b.6-)Wi$-bSF'"U#!,"81?g%.dW1+цo^3 )i\.)!-Pu:%id"%KAX2 GQ]5{4N\W*\h)G#jϷ `$rCBi4l2O0{kC()E 2ИJC*}"4*ޑk Jc{PvNc.P瞓$d фY)M;E!s$+!L[XJw[x4m!txݲ,Zt< ҧVpc4)@2S< dtP5hr V:^(CtPcm_$蛓Nb`ZZrTaZ‰dJrH (.3h>-?V[JT:N-­Ǹ1: rFl F]49-IN94䟣 ZjFw6rFiX":Fj&mEĖeTVKl?ޓ=fwCh)%plP"وP"GaKgfP׉A {ȃw/E,,QU3WѭӓQ9c/(7 zN'Nῃu$[$;US4WSHGUUI{P/r((s=]O{XZf g#/U1$4xP i9=%"whw5Mzl3`/ws7{ŧ1.2g(fKo`r94Mk,:Ԡ^v;#9ҊsCM&pm97(N `=)x*/u j)ÍzpF9Qa0pgfJVWOTU ۻhI&<5Y𵩙c-.6?Oq;>]{V W}P\~<[0k;/Odz W.b0'kor|+6b0r:$k$_m=Y6$NuöwSF*#F m`~, ,WMΪ2ruInu>ןz⅓XfPer(L aQ8+!Q/ڟ0ܕDi8bYO'(|SRvP^י76*jվ],/__$ \2SBuDfܲ'q=4&E5qykGSU S E荎 B7/ӻ×o)3_>?pݧqt }!| .u_t啺-V}PhQ{Տs:'>hiJK0dRIZJo nhF,<}vU:a@X0A((Ꚕjs"NV VF+kFks9-צhqy˩γwYtv[qّz;؈g2>8Z269w(Wk7a4U[7s]4\1VR)uKK= uOs؈&";><# *8#rʼn& KeRъY)V6f2I6:AI kU.*- KC$Ec_'F("jUmpSù^G[F~/GmN+z1eHY)Ej< VBIPE䱵4{C~͂ _5cւVLWB]V pʭTO,ְ'TȗG0-F˚p=^ m)9ΪҺJFu)&nqԷ6WV5?e(Zu5ٵk t?3]mi1.h¢t7^-[(KC)8^z\(a?l'{O{GK-Lh/!rRnBaIp' *2AlKE*(4$i̊0 Lhr=[\"n<b,9A]` 56OP&'z!sQlP!p6>ɦ`F ,ePT4 $Evjd$h*h,o˼3fSxo)ʷ50x (z߶7"n\Aj/W<_QYd 3J<2DEJab,]jIDk]o!_rǏ/[0uD s5s0]SFK (@RsÉ Z0gd V0<`  !50 JA`cZIH24"F ibl% bˑ1yI&*as[mã㶌4lr4!DN Js!>܅eL(fX%L+x<8gi%Ϸ&y58IҜ@DH} ňraD)$ɓ LGk>$8O,9=G=KZ& QB\ ^eZQH)hK=I\~ WzZlٛ&CC Viͮ >-^jBg=t QN3Lq1Nc"XNhq:M\xDu(e>2J&<)YIb.pH VcB5BbT;OGeP>1'\h.N4x+V{Ll?,xcx;V~lT|[ ^@z0;o8p4k1> f$ۊkfX@&?i<<)zSFtG ;{\bwbYHdﺱCDimh㤆H\+odʹ5XjgIŸNzyΓPO',@IѳT@A0!0+.BJQ`saE\f˱TK ,9W&2Hk& O5Zks 2< 8C+ڥrgy^Egjg/Lovk&(׿ZBTۈ.:Bxm:i!LÂ,("NHٹ%XVcbu%msJJMd3B9x`OOm;QxMǝ,{U I] UU3/qMm"ٷUmw^}n~]ƟG[Vgl|q'-r!aפ ˨x6dzl3_׮9* ^9 ۪3z1J%/h#s0UeL8:z AYqE@^+y^I7.Ji )$;?6YkC<2ǹ#c3D|hh5 '3~NiVuC.E_z Q~uCW()1o% _³yQ 7b4 5eNPKΙV2+=OaY.5*yBmD!d1e k0g x gy菞~GO>ӝ9x~GZw`>1rNQӂ  (ς#ɰ@5\S8S\hd+'Eg7T\HDTLFVG#$K3J1&'[xg ވy<gi \^bWwQo +zfJTGC_n~g' [3SUhrFz-2ZuFQaGr (P3\Jd00MLԑ2/)8RyIxR.^ȲQhc"õJ肷 PF%S@~[#T¸8J; 뒌8:EQTR!VmeyߢQܒ#" `Ƿ5/loP%.=&tr1(O\/l5ɕwGE-㒄ck_knL'E5eڹYRl6O㇪^b6S)&k5Vzs ү1n8|[# jҟQAtmSWٵc0/pdn!elJ$6ib M {!=vkpb8שc_| w'`f`JIvI*mGkhpOĶcSȸM*_0Y62ޢ)4Je[i[(5}ʌmwy|w]pa,棒{v]%[^s>dWW]ڧh+pX~O Us˪3Z9i~W' ?d4XW EzzRrzzFN'IɈ&.}8- <]J;PjCo{K=@ G$h5MB G9IQ:\WIH\" CFdHYQ x"^"p7" ITGeK+&~{ؐyul"ӏc-(lEo{!!oNq &) `‚I  Ig(8?Lҙ_2" 75IRcPqy QfL£4[Q9moo?_E|thyNbZr]Tbowj5_u07@Z-LBgm7ȞԀICJvvla)8PgO` F"]Ap}+k+)efjLHŶ uG"Ấqu՚zn: Q48 eR~æ9_M(o v\&4jmcl_~3M*[j<% ޱ~1MC>gvi('{ؤK2C ^noOnWzdUhloڼbhr>D[0?Ke ^eKjIP&sjFƆT3]O[IW|]i*H7ݹҌ~~cl$V?0`}U] EŦDY;<=/j]F޼ԛ"% l4nFl# CE*;T4nWFI+^4R/ F)IRWʂngU%W팺vuUtzJ[vH]1ΨJQ*\mܫWղ;`-`gU%WLJmWWJz8c!u[vF]UruUEꊩdzu AjΨJۮ*+}z=ʡvA&X1+JԂvuU\+@NcWmLs~#]+2`%b72 h㕆Xb ڠI!8P 2]Ki:T;z|Xeoo&j#{ss|rr7?_)o6' @Ԗ-3wQ֋pmdC77Bf搗CǏR }8{|A.tAauB9+I}51oV2Ti!=xg}"Y$elcQHmY_ ZC6Ħ{Bu)mcJrI Vfi U }T|uF9_jZb|h {͢=.6rj?w<_:r$@0a mPЊE%B/<{]6i7gX䎤^tT8ncZ&ނoϗDԝ7, o PmscXbrVb}鵵m+6kD˫(Qyu Q%lHbXd.vBMI8eC .Zt@X%XZg  AJ9$c'uZ]w茜{tdMpt'px|CR6Evnyah ضnTrCKFgM`%R!9P@P5>׽v'\O{9790$;]#R 9h j,j <0X6!%z*`uedFv%g\)Α6F0̙ Y Oc9댜{Y+W_1[¨TTT#B "9)O+ *@D:JhoDEq"pM~a'NdZU1gm\qKPlAY~܁*azԊQo'vҎT6Eb#YY͂kOˀ=ҹ$ Ou~NFr DqLt[vVn|m:!b.)4NA6m]Q*g"륭;v~zz-lż fo8 Y/HXZ_~򗶮tld@joSm|=de;I1IjF8,P,.k;zG^.·4s(>5Ctp-K>ˬO.sXQ/*arc[^KwV7Go-g.߿oVit,x[+a~ ̧/{)3{yDъ쁅D Z9,pד \ stV~\'>4o&HV>nhWL;M*k9?ߜ)UWq99oGK3s_Ֆ-}\䗭TkOΧG7ޯ5Āvj9 e4<<˵]A '_Χfy\QBliVfH`֏tiyì2ڔ'6+XWb'c&GvJ>bF]FËǩ YUr*i]mJxyIͅ Cfz'|ˠ}z,pT6Kjqɬ\uSK]ް*$eWx=^݌[fT9sT-K¿gSgO~|~?;?HO4GF{H p&_"#1?aMǗm]j]>O[ ^}(^,,Hҫrb.d4$,)CZ3;Ѱ-4W{omȫD7-%lkJB*g/ՂdNIyN'9O|'!9k:k|2!;h RYl&ؚݩ_ɵ+e+eJ׷t{e-&4f~ ws|BmjR!CC2( fI?}a'kO[JlQ+n:L*V 'l-)/HʇDP+:WCgn2(-< h ӟ"Yw럳<_%ۭvٸ9БDpK,H9cj--LNG]$z kUStbk)IIGP3(c2ke|>n9;#7Z8<º%U9&rG0ȷh[^Jօ.?_el|Z&@@Gy%Ez}ਵªգe݉ngt]]l@f#BZFXGO7yՁ<=y&PCR(ZTJ6 FmF#ɜ0'MRJ۬mmT|N9@a D% 1Y'tQuFy׵֜-):fœy:r@8>z^o&C>N)=97+e>]$!֖A&ɒDL9}z"i0ْ6H߳%m~i!?Orzdѭ_/8eѯL\. } eEXaed-R$,!ԖwV6@S -1":9Ҡl#Z_vmIvF=3΂_M;  z`zݞpiX˲/6^+ "cы?_?4ƣUz=s~t0*k'Wҏeklu+g9".gq<4Hm6eY&z&ܴ ƨ:XZPgѸy-vkh-I,,$U$Aҋ?sk׽$ߤ5h`1vsκkR%J<$l;RCqXX&I@[G䄷Ovzv@¢6KKCjPђgbֻdEh 14 4y:NzW tlqҍY6OS,?i&rvڀƉmFf>1`?XEuFcA~j=k4-&MN{6l)*qM!|%,5"2 ⫷g|r֗$Ɯ,#ck~?m!4OJg1mzi3O6o-S͆Pe:x.CDZv,6sZ:wlfm [:-z/ ˄Ô1/fI]E&{Es"߾s&Bmݷi{nr)O~wR CW@S (l  Bi &XC3,_]n6ė/TЬW`x= tTD"e)o-8FEhx2BF4%m$YUuf٠ ͍A^@{~6A ³Ugbqmm5J?OI=i-I U2_cm%on?WaZJk+`HWq+BymweX oRvjUZ) 2!vIYY3ȅ3Nlj {I@v4R(x㩎F&` ¢uqhiATw.?2qnՖڳx7ٷ ׅlKƓC8Ib ` ëW a!O{S̟,VֿG?..cD{*:yMv0Kh|<~}?1g`k(!*\tU&ÂBwSqޡ(;7}Tw%KyXJ)!ꒁPĘb܁!$ s/e$J(T.0IY]tG=y9$8 CD)b:#UP ΢^c=#&1 Gp &ӥNO!5b7iJM0( 0P0U/=R|ǍCB+E*]s9NZʜ<6g *.[3s$u,Yg19uJƉz/ aEBR29%кINѼk IT4)H `'y-6F \(| x@h9D"psi Qx85=}]9@,mdIHF{~2}B,v{n~*)Y>?ZT} ϩKi;\,#se, |*C'88eOi\ VG|{K!Ev@N'"R .iYXATGMurFBb0/:Dh;"i䮜hpȠB%NYN,C9 b8@֏k@w-`S { 2Dp_w*UnM.u %//Y nc6fKLJmw2Ĺ=PipB$/TWF^p9HQhTJxb*Ts^0\6D|[ʯأ佛}t6-(tȝ "D(ă.F+~ jƜjX쥮?C^m6w7U >_aobׄkp~;}E7cpysd~xƽ\n}߾&{&oC|/ϋ츻EZy^&=.Vll/wPDw# IV(\ GT#Ȅ=Օ"YZP: kp!!޸( )@ǝLM9q.)hHZtP9Xύ$2& *HRW{ZĜS)Ux8BW+ 3Sގܑ`R/?:sw"~ >:J!M%qio7怲c;/>o=?|{3, ҄2*3eV~v𳃟g ? Wf&STBb*2H I@Ec kT7byql;M>Tv6ntq/G^.δ16 &(GaD!p8hɂ+'E5 )5( o5ELhRܢRٱ585x#x<M6?ų7_{ual3^1WbJU|+`{<؊(*tR<9~qVsr>(YNppA* UKi 5:R%NZ*/ O Y9m\;MM.x뜡B(Q"(րʱ7sSήqrgfK2bGEFXPVmpoqQKz~þCYj .ϣ~QKϤ,)6clM^-m<zPL̅}1q7kXC>)qµ6ѫCLZDOxD5O5$ϥ׶;2hClJ"վCluϸ ƝqBt٬Yns/sIiVcix w'4Fm̮MmMws7&s%*ܓII^{hܲ_0YV[to5!w6eP՝ʹfb62c۝~<i}ko>Joսj~(6u5/=VOwo>9{f꞊۝{.zǝԜ7^лsyxn'Ғ.nZއv^isH4xNO׈M$i9%iR3KOH]b^2CdcO8UPNëwsm=#GӦh(߁^6ʀ_H*9UGK>́ RAy1 (CWCJSDʴIsgtNBD(,ZZG=ip;œ2UIi!"(Pt@sYbY ۷CIw.prZc 蠁z .F1}`:É.Qe7G%@>B÷2O}D1\kA^ѾgƫLhTC@b䨀JNӍE~c@qS2&Ǽ4;블>6#'!RD 1B(ce#!CrER8ehz9dYHƱ`c) b|R &4R-c1q[rX-,63-d-<-\9t22ޒyx9~k$;)P`yПLtI$lG&1a02WtTRn4 Klvqv`wx/D G$hNxkTps,!X0MN9ir]U& E !Ȣ= mMKc̼THGR-}I8aM铳a<Xl~kE"vq 'CB Nq+L0* 0 ֳTNI"t&qH%G$(θٻ޸n$WD14c˂F21D{Fđ iZ ^<>\+:ˇ27]8x ڍMpIlRǷ'mf袜zr*gOLU'N%&|?n/o>"E"j<s2ЮG`n@+wzc)Hktt7zS-mн w߈MܶZw #/hט] #;* e!;qXa!(o_gpYl1&Еa+tesK8td!]ז p] \BW6C+AyK*]= e`fJ2ozt%(Iϑuw][y*]sY +Aӕt *L] ఝbPͬ] Z>bPPt m}VJW8"^BW<] J-%]e)xCt%Gkz>:] [ JW[xϡ/{Ճ׮7ezZoҾ**]=zsV+d[˂ɉ~r o`rOq(*r ~=.Ϗ ?I-wڝ8;r}ڪPfwOkحZkm|gИ}o}нOkit>pmsaGч]PCfr' uC<qqaitA|Oes ތRf;*9rs\ޛ#JO-$`6T0m3ՙ%LxՙLIgX9+Z ] \ćNW2:gHWS-} uBWcJWϐȧtf+VJЦtt%(s@'!`n3t%pn6Aճ+)oon]7OWnP'|>wRPpw׵DJ?||r"-tmY!ۿO/ZYocT-4..o1Rͫ_v B4?paw+oJ;|e?^`Kf4ψ |[o]DLՋ>jQ ~>w5a|&._Gy"q~D, WG?o>?_k|i؟#~ͣ8ɭxa#^X3EoGbO/}Cy{׻kHG $PWkf`wǧ}.f͘lgϚzb/W` lΐ!sp=ה]-,ozɢ miޛ r\HJݵxa4鱴~+ζy42X~\hD;1MG@1=%K"w ŕiMͅ!n- )6-pe%_{`d :44]-ٗjLEZ9{zh ׯߞBR-Ҝ!^icI mUnur0f9=b.=Lf t.ZD34 -fuj<[-S=m̔QF>ׯ/4ſۄGF'9*xr:ølC1iRdm`1/}NN (sh )llN9" A u dz$i.?7'l8qiֵ٢M \,k k;(c?o.B2MfIʘ) 31fzY*F̶gl^(yM>lن-9l-b2<9si!s`@f3e mFi\aM&VCJI v{JЗɍ*s>7XkO=N`1l~ռ|kT6!93M&]zy;aYgJB_( _ˁb&0`)Ug`򀢳|R4Z ď&]cYqppOCR:zCXcć1Hmm B˵^.jZf} Iiݘ,w`ZXS0:TkƍCٻѡ*_,%x_b&؉~rr95S{!R- 5 #̆ e@l#d`AV#f jgalW(M  w!& sdzd#w(!Ȯ܁vBnTzCţՎd:|AuPe0BJ_7#Jjk%ŃZ2C6a1%G$V".b$) >@ E&rZ'd^0P>$O\LR3PpTʶxPI#6H܉m U]ZUBN ~d}gnd*]0\ l.$`]0Kv=NN1y SB,dUW1d@@FC+|uiS9,yj"!I-G]%@_)W*7aDteP_pή%Bm9 oюa<"[ހBBKB.%mê1$s'֦Y)a8%,@GF}Zl:L/dM3PZcCAČMDED1n &9fU\4Š|;,⬒)ؾh{jj-[` _ket՞Ew4$T52kE zizСĺ 6a|^yM/]/Wi/N FLJcA.naK`3^o#'Mؗ]LiVn-ɳٺ bfIh^PAٰW'=<!2#r,k80)QC^"'$ڑQ5r#cEtNJdVFt %(H H,@z|b `uA8V W XWHԇ-9i upp#.r)E^N Q1b0S9)j ꠀ:Jh^:H5-йg.T:a+j hNp>Efn5 d@jփ*Ú>+|c<3r T ބPmG1~]Z]-kPtݸ<^0~n N衫7 M}EOpUڬےv|7nXAXdgJ>J@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N?(I>ߒHmX @@yNgB|ZN uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R'uߒs؎)8Ńw;(v.H@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': l@Q!''M'Н@@o] NgB9DN uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N u?{ܶ u6~QUjw+NPtu[ڴEa 7̜6B +۶nJ[QЉesc֒;^1 d _\*j/B -Kevu3ZUMa`bO-Q ֠tOhm3)8Ϊº .10!sFozz*LJ ƍ%,{pRg#z~FHTBb /-2d:gL(3! + j v QZ{g;cS%:DW.L0 JvBk!^]q8]`IMg rBJ=]Bֶи+gdRk+`u . XWڶ伧HW(թD(IwbWӮUt(DWIWXKjGvʠ֯ "ʶЕ.+TW vBtЕsBo;9yapOwuZ{" ,@WJnW 1b!K;>в)`b+h{&wWҶOZ^jxs6_MWt@:0m;1A k`j^hEs-U.\B7uk  L\oՇ|v7]Y\}Fr}Q]ME('S}5j?G ?SPׅ~Jt/_$V"'UoycЍHUᶛ7,~0ތyWxcP~N|BoFo _Þ?Oo>$q8u{8IJIV+Zbrq~.+CBHLd9r"/D<(gA,qy/ZfW"I.n)@?KǤxY̩ e0Rfs%gx&q.{ZkO ǥd8vkG1.0|[>9] b~5YL˖Vތ`.lJN?נo'o|]rYO5Ftx;Uc (_[C p7˺|]C)P+>$¢P {# i9M~rޥ iQQ."I,,7TFܚ#0ěKk3Q wVe@`.qL28ll3995c2(@v\5xt2ruVrR~qYZvFS]|U]j;}} R&VCWߝħ_\:Le}2iFu~nڱվ``<)^ Fp:L> F-Jp8;[W ֲb#Tt9]@3ՂCGsr3+g4$bCвyL.`B g:3.Cq]b &~ A}FP})ϲ>/V69'QmraӅQDksU!x0{7馲2h Lϓ!L:-@M9=/a1Aa wzF? 7Yū7.5u+\0B(v| Ы=B۰: ?)ӻ[jmr/|d,h}PEϫU5`~Lw`(& Nt:SvkZyH|C4 XhL0 ̻} SRLnŚfX6`u^]8e{2XƮeUB\3ac_nA#}QQ#c@z2 nݎd5 \{Tw,]!q8ݣM"%^=^S]:7ȝ ݚUc xC (]KuCw'*] ެ*,ҺQͫOW>_ޣ浒C9[]yB~Z'*G"V燉m5 EEz3jLq̗.Hܳ#qk&H\Um-GNimHFQ縔ť#pR&2)[8Kia`Zj|{֥nh;pJ$0y(MH U9L6 3vcpfl?55ݔtڢa=k.BDy͋@XZdMYcE:i7Y0KՍaJ/,C*$<fpk*~vaFl ӌ2P{rԦXgcRr,/yQNLIV,!TKn4('Ɉy%x5|ؔtˇa>4 PXݝXw%ُ'~Ԃ}J5y&|ۍ>8eyI{Z / qm{%rVtMd=ɁGaٞИq{LmI=* %i|"$G%Fɘ4d0 Jt9D$>Q>piԚ(c$r \&+ ;9Z6gM[䕣NZ; C |}4kۆ;exh{VTIJ{`"'B sl LV;g"(σ#6ϕMy)@h1A}6M'6PFӡ{eiN$pg2)ΰ qJdLaDECU:[)6*UGͦQ?|Aiw K uIFM#F2kR{$e1'՛xE.xQHLg 6% G@H!e}? @A| 54[~c2.'</㲅^Ao&\@l᥷EfW~۔7D">l Uߏ8>%f\8sqUm'EoC;NۤMXV.)wJ"B8K6@sRddn9n=qu\oԇs}%ar34Mpi\GbL'mC_/o:BlmV#Co͏M[Ͷ4;(yd7B{ϼ1@cxקguxw5ߥ5[-݃ٽ2{S4#Pݸy5 A'eAe)D8"#H0eo_ / < ֧w?q$I_i #3d"qnnO_dC +""c֢D6"h J!R$R 1Գ 6[i R)(ZG[,Q "IR1ZXtu9XC743TݑWN6CGa#ϯG1 bJB>`L_JOn֩m[Jb򞢒Ϣ w7(:9\&PD< a}trgkKnQR!M XY&Ce( zD!f%%Y`[gMmMQ|6VbC$ +#zHDD 3 XBfC}6]sRdi-x4 Ӷ{1\m+yI`nq_;ħ(68}(Ww_C^RGR<+IQAH&sg$g|Zek bfTj³.&~C@T겍c&ںk+|HDIgKKʡ 64r/^Ʋq)2trTLz/3Vۤj8ƈ= ^ƶHƶ<@ uF1kVݵ^v.1Pr\k˝7=Y"YeLhE1%m >Y| E?\)gP1Q"R8` $ʢ%:,Y!"_Lxc/=Ӵ;ڬ^<{Yo{q%| m›;Q KѮBbR"5րˢQ^KںeE9ʲQސLhIdwh,&H+E`)&&( d踄6g-"P&}5|$^J+ =0vj3oUn69CC))Eu˰"\JE.W0Kl[oQ3D(:,ӑ{<[",)z}5>JPv'"Ip/g"CR }֔/Iڡv qJf~ƀvQ:lЭ2]+Yfd6Jxґ^(dGDR"xձuF 嬕/@ZF\Ũ|z[ҡJa.d Ƅd l!BXJB}F/wm/ cL!#_1Z#5J6X3f؂2#UBP3fN'ҠЁM(8gJ`eQ JD!:Qȓk3XYx POl^qt84;9ږĶ$d&gI65)*DeZ2.Xf-×)38!l/myDcjv+_g,r[჎o~⑜IB9BPPQ I D ئ'J.±Fwv(+0֤m8!;'Ӻ0ڍ:]_Ҵ4K΢ݟn5He@BlH&l~8֘e՟5aM\48 cשּׁ>35܊J ;"QZ e +! !Avl+mM.Vxٓ b /:߶!G圑D!f3bnJ<I%df rd3[M㤝w.o|l_y5$Պ pm |9y;[),d`\Ϳ}<iiK?ބ RVl`ѽ A'*<_ ~)}١ JQW&^x|rњ$z0\Epjmw&b"]^Po5?E! s$!U*@FDQR者,Չ0A:|4{.‡_`0oߴ>Kɚ/Z0RasYpvxtjʭﳓg<?¿gޑܩ}[\>$ΙdhzC^m?Q;c<"08ׄ9`{ CT/*azOl;֭=vwOawF rq7>bDٜ?;Jٕy֦$c{~kvtJqyg:upFY[,%"HxM:ZOxFn6L ٯJY`>)Bmde^uCdhZYW-yQң),~;#gheKiq`~y`T:?=dqqyT(V?2|ϭgs{5O?Zo5#F!6dƑPp0r0ojV«Ok.,-e=29N)+2jivX_ϓeϦԆiE]{[> ~:8?Λ,Zw~E8Ul +BKV^~.ytziSejX{ >oC?ǻ7?0Z,k IN8_ ׷?mzu[WFalV/j,SdTQd Rf.%i`+ɕZ;Ѱ-4YElyȫt#\2># \aׂdiV ~N'9O<PwZ0Rh5!['K/lv'y5--}|4#V7,s ~j?]-2:9,_=͟s `^d5"VbWwW;qfw>mˆ8/Q8fː(yHTA&(Qg4ID}[=R@P'.kfLĢ̥(1Aye1L&,1B YCR>PEX}A0AIs.oit]3r.kJ-t7ṢkJV<<g7ӗER3 s397ܲmFNx֖Iو]M(/RJzs#)>rK~C=o{V2c7KfS& U+`'BO}(Ơr E]kh+QXcUMlKȹq 9:]^h%?G>4Z]^T=Ν5>9' Bt I rڄM%W"#i>{p/`KR ٰDB A,S֪VaS{ •$e8,(}KNZcX3jNɧ7KMۭg096:(bidKv)BP 5iF皹˫0ъ (I ՠ5ugr<{ճQixR*UvD+Ps"K2g>Y3 BD5O'F.u#UWc5Ǔz(5@*+e 5)^;u(x; [$^d PuD}(ڲ .xEKNO'͞\DmE}tIa*)Eձu]|芒=xMLgB 0uڼj<ɻ(?j$ ,3q9WeaS!fzz>g_<rJ57:u)Qbf1F(F$#"oFb~5>f|~u+oD\kۛOoE︙[Xs3Gޮmu|[Hm06&Z]+e] 4j}2g~\dM4"L2gtv1`D`}_F{g[o9\7"S[?R%Zܵxl>PKu,p)S6׎ds: L{mVR5Kō_GGxY/oy'-ezwKPEY 0jcWHV}i l^>,[~*Wuۛ XmXwEY?XuYk7~:X}bh6JW(Pv+aM!^U~trk#sm_LwmpZˏo~UG#`o2 ~xi^3\W&>x=dzi"=CqDW|wz=_snbyO{ۛ O_Om*ˑ{fpMnn r=L̲:ֳ3?u7?Xytp}osmWjzg}Q|yrIRkI*%0YT"7_aag |BeY^]6ҳx!:ziﵡJAE)e^H]e:OSjehi& xe*A-%U$ٍ̫ijQjDO@;FA; *o+ZfK?p0W~R̓)OZK!Dyҵ'ҼooD}b;H$̨"16MQaUe8NeX eX0zψjO*Vy 2Z)!+bd,5i ϒx"%B&²Z?'2de-<:*Zxa2V,7 x\Nʲ*K3i LV9ϕ*:TI*]xdR~[1UU)V2ןb `EI)+ ^U0*h0MsPk$T.]%k7Lt:uCޫNLqjNژ]o{Ơ.LO. 3\@`Am0 ".'~"R*@Sl@r%9&\Fx(rL.WHuʕЀ *\!޺JCIʕ\in81`AUL.W@r5@2I+B+ĥj\!{WC++Zt=^7ƣ\MΠ+ʕ3n{Wδ+gƊտ\!U[^ꐮPr6/  繾77ꠍu ]`#"'O|l#JjBɄ1B(bfj:J>ZE6 kb dR?Q e4.Eoclizy=_]xz?%pKx_ rQc4[s7PhU]ӽ.Ïq4kokg tW )mEyԪĦE`zXZ9B_1L9m'igL Fr>+MSaU.M.OI'I*eg;d՝ vs(W5/)PB\$ͬrr|}ݶKE4ϲb%K'EbIn KeiHzMl8+چbc m_;8Nʷhc`2%+,qy0r.WHiP\%R\!-޺BJI\ PN\! ƺBZ}+(W+i ؄#Wky(rQ2jr &$gRj\!Pjʢ\ P 9o՝\" e`AwBJjre&@<B\aB+wBJc\b- %\ : wܕ9ۍVFi< d`rŢ\P(QSl7|&0[g|}.fshqk2֞cIԍ,؝˶{<wgH;y{gH˄R zgh&0a?r$- 5e~/?__ys(E(^Dde")<%!N/~~[y̗|.VGt17˅2YZ&2X 2 $Ԗ\{`8rjN֋߃i~ʥo3[d 7iݓ~}Vy=V MɨXL?asKKAc9ͫo𐺛On^GsG:|̝7\|Դk Ƿ.BjWBO+cC4J֜\ˬ9٦8I76'CiX.m~@˲e '1W1K~|8_h}.@﨡w/Ja 6x:mԊ_IVt*n{Vv[&dϮ;_Y lR.ޅ$,l 4$4Pd!mT4f{t*%X!i!.5ޣ;HCiE8xΜ. p=J%Hg~M~%>_q [ZDNJN0I3 z2GB ,ONK)cm>7g{tm(7n] wF-VmW+ xncz&UĪb*L `b{ RRc!k\!nŠh%^(W+$ `T0r}2wUwBJ\ P$uV+f" q5 EuN(W+Ō JXwg-u\awBJ Q4ƃS-!UkB+=ur<Q ڄ׮ qM0{W@kO1x\![C+Y ɺ`)" ioR(WFxˮc"2pp+7Zݓ3Fi_rG:)Q!ק2fDDcNfZ^e:o_KHZLm}VH^!O`At06`l w)6m & 16 `*‘+U$ZFr,r%ќ$Wl( F#WH+rE\/hDHlI8rT"WH˽RD\I`ən}gιu.JI H8` iO|p\եIʕVLAV2\!>) (5Hؐp q d@Z3G\Y5 ɺB`+=z\ׂ )yt_\]/DnϜs2*n7n'r4Qr%\!S2ng0vWX^?_&E u|0x?,<$,s-,)'wFѵK$䩪[[]{1tpbѦgOW@i/ `S 2\K+F}1o-trʓ򴤵+ .]g7BW/YrKZbO~MZT7y$M_"]E~Ke?nPj1S*HWI9RiAtŀi9nR&*^E]&;g?E:e:o(cVc^~:>9vۚGFӶ޽w탂:>'ܟO\io;迷Gqdm;޻&~ןut @o_5''g5or97W7hmѸ|ܜ] Xos8öiCqxo]<ͧo7h~1>_= _)}/M}>;{̔qvYھ7oR3#PQüܠz:{?[& o|D|kAm8Sc'ݘC. _uߞGFِ?*g5[k(}-nh_Vz>fC2xY7h/ߏ/\>oe}|p]fd]Jz\vAɚB:+NC.+HG}l%$Srq:vZ*A u֪r!C!Lm v#s}atlMƢ՗;q\C5JsB@MP4*ڜ:Lu)noWDzϽ#BMTDzo%[Z`$\h) mGylNT|(OEFu z)$T,AT[1ևAJk\3: kJzpv+5RœDhɋ=b.&3q5sP }cK5cE紫Ө%gC wDF0{}mɺM(Km ߕXDieѥ Faߧw@>R;lnc0j!saMz7'ߓֈs.Ǟ1Ǒ֑ ~F_o3dT&dB9*_*X )%$ڨf]tT]X[76Вjn,ΚFZQW3j(UÆV9[0f %xDMU.Y{V+= RAvT5JZGviENeP4BZ%;6Ɂ)BAnQx4{4:<<5ӊ#qI0kD(Q]鍠v]/ ·? TA8&_XLĭd XqkfCX1-83R뾃mQp=`$yc#6JQjPѦ"jB-r  \: RL1 v[jVَp5TXvU s&E0W  MG: ŨQA"R}W:a _@ezm|#"MF2mv^Eo{jPB]ɒ܁E7 +Z9q2FM[@zNB€(!2iNwQU0uk)& A'b΂G ݄b jS]w \t͇*QM%|7 d&/sAGZ(SѝD]`#)#MEUPn֞u*%d@BYAPSWPci[]TE"YUDI)bugO(`3Xo ;HH+.`5R8(l,@'@HM2+V*{M23z[t{^,KT1K=IkB(}Gء3ĺ\]nz{|~׷Z bcn.0QCیZI|tx9 ^%y:d6 jmlAhqU +0똆'z$;R@f .w(z+R|$Q*LyM,CŸ04X0Ft/1 JWLd:[nmG⭐C@8U]UBNu~d}:MOU;Rݙ@vTVW QHk]Ư_sͻךcfԌC|#S$]:Ԇv~sߑg]䍡"UA .>Ca E5FHJw:T^:H9-й ᧀFS8&գ-19mnmVܢܴR3k֬UVm(|ΤL BVcږtz] %'ި \K}4 ŢWqdMAi "Y'XkʛrlCiAy/FF̆ M F:,`=YJF$=kikJ’ # @:"@.* 1[.ՂpҘ"&b9R Y 5߭8I;.2 @,'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q\'GP%9@Q4߰'Ay}*@S 1([O_}KR2Ok؛ε$ AɵCrm\o_۟3ݦ^?nq5oh@trVm"K]q+EGd]¿fwhCoX7k1(ސQ6ݏnQhSZ M3Bӌ6si$_ M[ŗrY]X}tpZ ]1Z;]1JUibZ]0+så ^"]vҌtpbAFk=]1JBW/W +6:.5K+FKuTCWh3lrŜeJU4N CEW ٥ * 8.gŬ]km#E.X| ,zS7. 6) l|BdɫH#ɖlɚR<ݮמᐇy| j=1q2‰*E3(=xuLGѠ1D4$Rp]ؐ|ɂM3Jڙgfz:v- (@rQy6D{MNg!."T'p™"SQeP y*F!BS*O1S1,㤕|MU/lΤ/gDp@9U q|\dN&9N'>2 8{"=Żo2=exbx=Y_}M.:g#;y8MK;9WjK?.t?^a!DXޟ;uI !6@0/^ܑjw\6M!r# k@M` {.DeoYC U)Um zc`ﯝͷqpJsp?̊?t>]Ag23d+tVA.~t0@Ws`y]* /_0*-U1i/fƕ"2)_n'ۛR=ԙ3|1/9#z؟^_rwu7u#wWMsi}{pw<W`)b§ʨ:W3M⧣q::@4_>Ӷ3B*CK5;uf{V4E` }y|wg:Sv2<& 6׽ѹyH(Vt1\O] QX2N'?P2Mڮ㨗n!i tQŽ+xaf&N:hYeO-A:G/cOuԜhfo:e{@InRoVS~썣GeUEY-zev-@ײm!-!zHkeRkף/VL8B 3[|_=-l-WHz8jFb@$&('&T{ce}@h) BZFRQWŞ(*H\@D 1͐gL l"n,Yֳ(ɹ ԉZ d/orGeo!UVA}(}g01(rGjW\̀dЫVܮ&cM`_^-#EM^ ("uIPhKB   "ZmDxg{݉zSd]EV4=ܬf m'opNK%f`&Wc*J}`Rvs쩛Ktì;<],F;m7,*;܋B Zȩ2IhXڙ8hOv.lx*և5y*Z6Ԇ8ݣ"K^ ҤEwS"!4ZaGQ4=C8($MtQ5ꜢKx/k ]1q]6 &eF2 ,Z#=e-a1rvNTgQ"n1'Ua Kꅩ if+S5 !>LPZWWZYYݽAorHLQsK7X>5DU\X(!qFjO۸>n>nzq'o^*7WcVw ;^v,֧K++r2Dw&ڄň 5K-rl6CCɪ$ ]( yል@PY_ZayTpJƣz$-£Bs Ͻ,']1$ IhL.4Gˮt"aV6 B R4 mޫԑh\_`(UN+z0k]֡A&<.ڗ^B7R"a(BojKД'642Oo>z?XeXWBɱ(4F$% wI k Ƈ`;# Vrߟr!$E>Z&jN\xLj `َ$p!(/mv7%( #GFH+$02Kh 9Q;H97,gr;Q1*0#8dyI<@)J 2eu2@#>T VhEI>$=OXEFш+PJ2#*)S!Aу8*"Z~'~R-\GLE((h&)ȨR*mPxr(5$kG#㠖_.!x3sT<(]1iAV!k&y*1h>Dj6"olj5">ڰͷl!^KOFޚGFѧ .rV$Q# ~xus(oHh7zyWgqü~*x{MWƈ ͼӜn9M' cewF;0B\s%9ܟ8^xv$M1{7$Z@M8=l$F:dAIT 1.*he.Ke==i'qf-JyV cy뮍S4hhdKP;+}Ts,͏勗2]C p 51sojlYR 3~NJrWn?~B6Te]-]njmnJBefyB`X<, @m.{:gVy'7:?o PQeR7YθJŹ ?odԥ'jk+PZԕrAGosVϫ ]^/,$h4\8a~V]PQ=]( T;wuB'T^W盗?|e>%™?O 7`@Mo{i_glhzey.gDnH#DF`jN-sVK#Q ^N!$ad9O xh恆Izρoy4h4F+k/M#$S$I*FmsY9-Mw1:G=V,1ݲ4[vf>u,|bVaw684!TP[]ac`,DvѷxuȐ"CM<[ rɉ"2J'FaRu@8ȥJ=I2%( 2\@I di"tD, LцA_c< հ7EvE>6]~ZO@[YݓAt+M\d8xBJ@"dcJ1:d\١5`w=}k &7{[]*ՂSm?=w̗;2<36<nq\|}pkL { ق]{} ٩濫 ?}qTAU }7>nM/|BdiyܜnҝY omWS.z70Ob[܀y?v_X /daZja:ַz[,%@oM$k%ǚ7JPWeflki،Fh#v~zvЏ/B_}*s|x CPMJ1V ,#럾6π:r"pJf8M?&?UM0_9}e+}=Is㇭m/ X íiّ67Ϲܝ.;"fۋ/lR?ԒKY=hqJ?C9:&*.$ M|si;ɲF-5ysmOn9raxlZ ݁, 4:"bO6f :>̿4F]Y-5XϵF 9RR"գ^]Cc @T$S뮬NqVA h렍qǪU]~+EI+:x]hOݼmnRϡiH cgH.e:(d(8lHz-*F~ߖagRPLet{\ )?+E+Sh p&jJTC(DB>[$* ]0 o6.!n3sNI5Q3W%S^˜kszM<.T CS/ ,߫2|ŲpD~UYÀW_j /XGq;g9kIݯC +}Uug]Yd"?h\JsZ?R]=]ȞeK{ T.ZL$UKBʪאsYk S( (dmθРtq+R?mb՛Vxŏ߿*sQ`p.L"ȱ+3"1_&_F S\Zm֚Ek+y3@ "M]+wκa2pWxӞ~8ë*ˁb0w?K.S o;3tje)타 ,4v6yTop5[v=_+s5ME4H)Eȥs4EQ[2ֱ9!M.|B*6zW "{WASRqɩ;*o2YiuL,ZGfJXجg uXXxqᥕdNJ$%K>hOh4y=^ؘ9kVQy}IElP&Fk!PGJM#~ ERB#*j%k2U\L&ؒr9#6NƴF1&jFm8j!jPOƐ(:3ACbgZ'Ua@-s]4 E 4̢Ύc e'1W@ HLq'14(HȔJaBJY-)f_ˡdit$,& ^s2,WEC2£֐-؋;$d 5nfΉhz5r5ѪjCxp'?|gvy8-Y 5wh뺙,Zh)IRVQ] .y]ia@Uma=3_"Y) !3r>Ġ=S6H6'}o̜ou w}YYDU75ؙ}NBT2TZ r^'ٸ4J]BT}j߽\wA`&cc6쾵tCd%@рC@ffE {;]_q%S:YSD)U{#2[X:0=3L̽\];s˗;RbiD[?ZvИ ۻK%eiJg(+? ]63'RU:WXuɾs[阆D^XdL 2k d(O"Mg<tvY}PcHv\{8qD}|3Qnϫ3 {f!w{ǓQm~M{]~c (V6k:n-cf=moB.]e5gNJPςE4i!yzLuOw󊸞zS<<~G.nd-#S-=/=;W~λ@1j7n9)`߄'}LmE5~duj~Û/ެ!_~rN1o.fBW0\^~_ yi7k'sJˬjk~[88՛*{v|7߭ڒGu>v5|Bmu78v3D kݔk>f_7Շׯ|꯿ຢ_:˫r%9}o\-J.ьo3:'ٛfg_-6l\z=Y=u*^/}|)zѦ.bpZct1so9-]Pԑ@׵]< kMYߢA@4. _]BCʜ bqq y}tvn꘽Ҳ=Co/ė[K-hegJuJFm0AU+mVOtN0#خ UE V2e(cWPuе%xhkKH;O˓<|gxq~~6yM[zwbuw"4U}|;.hZږƇ׍|1jZZ&æ' ]iKY;Wkb126S{*9N$c+Tk0UZ:[B1 bI`-D["UVR˃;;{ޱjK߯ȯntqe?{\[..޷'ey76kywuӢl=jhcYc,;N@ߟ=-΃(..ξ`:_?֧e{YԦK }(>GnUg!(}ly}u< ոzv%am) {H$cV35dqq|MCi}:oPnA}nxD Emsq1i]]=/ٰ_ sVWlWA}Gx־l7|xig1^Q֞t;N@ދč\?Gi3*M9#`r15R1O~SFMpʙޡ$]1pbtE^E)bZcRS+ضGεf@+gˆ-HǼ-?_\ ζo]>2|FWةBk]exxS1[If$.}TMCjy;MuMpEh[[mVFel]r}kram۳l>i>rkf?Wsu 4j6έ+=f/痫C{lP`b̕?DRݏ/w4APG)3!Qƭew6F^V-HW]11RtŴ.$zu5]UtъZ%FWL&u]1eb^FWVh%=_)9]1֩)"֬)+%gqA>uŔ&]MQW^Eg ]+E-EWLk 2ͺ Ʈ99cWhJ Sbj"8%Md `k]13vŴަ+yYp F?]s<t5CZV=(ܞi3֪%n kqƘn@rAyUmݜ`*:l03?xk:,mє9v3Ĺsڸ{nrNagC91j-%`ZR12cL0@t/1tEz1b\iGچiS6jCԂtEт]1]1O^WLue4銀 btŸ6Hz2jzmX㎽e-uŔg]MPWN ]{Oށ^)5f]MPW(Q881b}qٓ+dp FktEfpA'K2Ohj+V78)"Z=Vƃaף+ܳw|K`AN=a.ei0ʘؓA+̺zlՃ¸ )V[9SOo2`y\ԗr9 ZլKdaL ذOi这¹sC~; (E&cp_4A35:SZ{g읡Tǂtq+ʦ+u5A]iYꊁ+µ*HbbJ)88T GWV@bJYWԕuQJ㢘 Ӛҙ 9 {%FW+}"Ju}Q˙ɸA̼M*'L'A+ʊん+)M NQW0Ay5Z׉ 2mH^WLstztz}| qAю=jyacjeL,t[POJ/0/:OjH'.E70G?W7v.EP&A1B#&`\-&a1S:1c +HWhJ;1b\i]]".j!6\+{ Z.u]1ejKQ^DW&uZ؁]1ɅѦ?̔=˺z]h^ FWk]1)C6+ FWkPAbʭYWUPࣤ ojL]W+giO]WLsgp"JjgqvݞRɒGWfϪ7GJ@t5%kFjFkGFuZ2te[nEW)&hVדc2:PL\`Ab s ZԲ[\)&hrq~kVŘŹRGZ[U1 T{#wƸQIk0Jw6o ]pprtŸH[}bJYWԕƾ5'B Õ3ʹ.+1j2Sf>xip-HӺ;Lښ㬫ѕE흤 Dާj1A:u]1cWSԕC! 9bh &+Թ38E]y JW ,HW]-BbJ6j+vJhuŔ[+ +vr7X)"Z )^Uo<}{a~ hHQƮ]٬V=/RFJ.VVE' i깆QG4srfR#}M;%&qi{&1S./c1ڣ ]=0\miK]WLluw {q+u)ʺUFEAbI JM]WL yy@ tE(HW]L]WLYWSԕ {+ b-*|uNRgdqi=+ :j NhA"`jbtŸiucWLi\u5t^7j'Z ϻbJ YWFWnϪwGm08AFF;6*(SKe]=ALV`>D|RnϥA}BTq+ebYAԲVs-'Oj}&`R~z8+E|*/w61)3kAN;`:65À+ btŴ1y]̺[A"`^*)bZoSSu5A]$EW l]13RtŴϽp]1eue2 vrf0؛\R)ꊢjYA~-hJWLbZ|tŔ!OQW>(jަqi{uŔ&u8 2n3vEZ%d)Ff]MGW1:o$EW ̻bܱ  c-FTYWYH-tt] /c;hH tg]=Ai)|R~E'}jDN>I!/EYW)޵erc0sƱRLm_cHVAI"j]V( f]MAWtl6M!vj|lu/''\b2Ѫ3ѕTtQ^ξ~(OEm,h6 juwQܲ/+] 򟷟-]!}E/+ON)~. ASS|{ (O@ttXwsi_!tCQ:^]EY_qo}BX}_՟|]%.}~y9xjgJYn=[fV~x]SWWkaqWݼc/6/,:G  'euf!!<qnȧu)kGTwS*#(zZGgnFDP.ۢi?;/H?}ͫؖhBRkknMi \Ӫjjg;Jp5K!xϿ]^ή['%܏md՟ֿT Eq|ִ^aWBT5ѴُkM]4ƷjAHub=5o|;[b٩ |ִMMe. Ikb]Ɍ{JZ߶eR ŠB mCʈ`L Hh@w ]UF]F[9_IK6ڐӫ3 vn*wV(olK`uEmSJu15IMKSs/vB,]Tjcvv΍j캦΀M*K|Z4Pc6nESֱBjSiv^{PWsԚ֡D ߧ gZ{8_!f |ȇ 6~$hB8sئlY-luu\4 4fթ|/dC-ïDh#!Q"^E4H/mrJuE!hi/Nzü 9Yr6 |}1>5AfU^܍S5ԺVDC*)IV:ѝ(sRB 1-0%(mGZdjڤТ-)$qu$~JaECh#JK|S.RHU}i! !kmMrzE5b>I'yRt0dA%pMNԵHcfC A0iJ!B QȎўD }m.5ˎ0H%GxH ָhS5>#q).tlhtec]Q?s%xYPUS! 둔YlXkYc51*T*JvdZ2ɷV&J9)O-` t[ ҀRS̨l`WP&4$z xqJ +Q5 7!:Ci*dL'BXp6L:@@=-V(!Ȯ܁pYV ++5 e2a=o-=P,Lh=4v#]7cEfJWukʃ AŜI' sG !.A f|)*)ԙ5TDqť#3إ~.V!joSCAV(J892rAj LڳFxwD"=dH_($6d^i;VW.#{/AQ*{RQRyklDzFӘy}W@BM6y-d9(-%VSMȲZPҰ*4]AՊXZI LzZV|GKZmz1#.EV̺HN**&6/ I;L')Q"_ {0K7u&~1պ]9S ?Ԃ;Uе3[Xw ڦL[oT^:ҧMJM$W{HV*U2P (yCs 0X:@¤S ) >@E&rZ5ȼ`|ڄLkt0X1F4/! OݗUd 2inuG-A@8_UTBN5~d}%TU;+Qn;"($'Mm2X4k`!g0Eʟt7(V1Z8RT XQTbwp\:%X:W (]IڈJ5(Z54)hc37 sXHkҬUg(jtϤyfj)څ'е=Dv *UPxZ*;k*އY[Tڠ@aI'X©rltEiaӨ/zBL iq#f0:zhceaPkϔ<3pU @r1.;~"P6.>I'Pa'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vȝaNNhqtq>cR zN =nlb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v:L v Dpl@6cw( @y@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N'Uݜ@\#T .( zBN hM N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@z:N7+S`W.N_VSjon 6eu}~ ZKl|Kc\`޸Ѳq)^z(j{gCW7*fS+뤈qFteis>pU̅n9RJTLWOˆ3jB\)fCWVc+B<ۧHW݌ O1Hp ]xo{1t*x O1Hpv~b (LWOfV O1Hp ]Z{w cUxϡe^LWy#8tu?Ȗ=*0]}GڜKEYL %`K_r^]7mE2O_S֗jsWߜ,ʫVcl%КvAے.U}-7/n~*}e?=zP~vc~HHWmٷVz޾xk6o;[?ZY/hɳ~d;/.[}HU ׫A$ gAv1NoGcL*4#d1tոZdswZ)eo/>tu7^Z-Xt o3/՟Noco~_TݛonͶ&v9nf{M4\O k1c81-EͫWc/SK@ď mUxkhoM_j8ʏ.y=Q4Fu䉫ղ43W& fcz16'QC#qer󫟑!OW?|ovVNE}4 8}bZ膘j עYCF}n8+"mCWd2*P_s|} ǘ_vkݴ-zFsl 46=q~žY}&dqZoZ1K3vr~!̙V{g kyowp<н`qz)k\Ǿ8E(/N=)LvFtΆUs+BP\eztCfϤt5>̆oڿ+B ҕJ]Wa6tEplujD ]Hr̈O1Hp ]/ eduE 1= $ڟ"]!23+vEp ]:`跭J% ] nNkW ]]Z%/*S!vu/їW?uq~(Ñ=.4ރ"Շ=-bоK'ҙfX_\{篷KK.84lݱz+KCWM F:C>JEepBjMQ`lr-}?JPms\rzFGfE gҞ)ꢱbn:EvvM1{s 3[qJkHu3)rn]GWm?/ӺpsoNƣ^uצÕd$%XSlrqHQ!HmRY4_ 3ejѿz2ayM+.cY{/ϱt|=ٍJo`+6v,w<Oqv录WgM{J_=#z' 6 TyG޳{G /&*bRDIVٍ>,H+%rL^@PSA(S™ K@q1BkYOUQƌ*@E XP_Po}fkO0ٗ.N{M@ߎ{9>ɟ+ApaQdlX+뛽6P68eØrj:ޥJj}V~dUܙ'S9ڞ5߲ɶ\~*i Hr-U9hH^mfD2EH̅)!Rf4dKJ%Qk7T@*V=9F6T9Q)mNFRjTk،Vi qƱB5gՅk-+2>NbR4=><..9O{!Z #5RK`Iy*ҚRm@vJ@\)TcAA^@U Fë\W|BљsRc݌;L'i<݊;Xk㠵G$ dvF0\ X mAAEM2PA>,X]DJd/:BkRѨ( RIDv>lF~};ӛlCj8} *l }"_6ro4Ml_2.K`yuu=yEb:xy'1aI<4Rrzb$)èw95cq/[}]Wg',|Je _w[ e W)u|s6vSfg-͗F&U1aŨODݖZ 2RPҾW==!4~Ecě-3[Ѕ c;THPKlN89H59PI<ݙxjGbQiPHJ,DIҔ`KJZ)2L5e61 ]vHAJTJ$@ E (lRޅd,ɉ7#|ŵ6=ĵ^vJ8vl\xǹvl^ޏvIJv3^M}uL0ykcV*,& JL EPo )t{^SIw@척@E2SJ6Bz?ᮞ唄E6g;3(Du)JR蔃 R$&- ,!,+75X9Y̦yA9,xfc&Yb}Fd bQPjjipakcfM'?^twf bBDz?<={3c!0J9[sqCz)ǔD" ik.o\U6SAOz[;~ݭopxz[:wڸ٠Վ11ym!a$Dd2]X :vؘ;PNd P c?FM8Hyt1V >h㪱6!"}އFף_>|L?Mv`J[NۂW@zRsvȣ W3myrS6-Afݻfw,==;PSrs{+&/Rw;*E):%ug\D(")fa\Vv}dt!%ؘ)/ FhNi´>9u:WP j(tD ɿV߱iS.#p0w^:;/)RsF'lD_bTRJb^x&[/'/a$:0vEJ?yS{ ʔLA% S#٬P$B f@d EP Bʾ@iG=A/590c|f_ɖj6Ym7ۭ ضB滌׋嘕{itW1%-GJQ SI$ueAT9W_T F0Mg@Oɻ V) x*V;K%4#goY߁ϦEٕfk;7ݧX$W kyXrR ƷQ[rR}OrR:]Nr<>>^ ]QNA-BB?j[qj.}Bm=iZ$I*Z+\pw^˔Fbd5ڢ uմ t m=9FPLէ?3B a%9ϤJӞ΁R9O"؂n)gB~^N~u&Wc6:{s{=g@Us;;e9b-gW<6+S[#j;^8YDB/7|cm̾Wղ,]4d`LUc|ca~Aү v+N8%Ix&kkc]fjdvKX-G ВLHjٰ|yӸ߶s:yen7+eM5Ev5P|PS)^N^_nWqGIOYtÛݕQǧ~ ɘb2xtutIc^O#;y[i[ޯfUမ7F>5{۫(!<P[{⹚|īV3'o#G׬y{C=j}Z?҇aU qac6vO/~}W'{jYH(0$re!,Eſ\2INiSV>PΪ>>Ns>]I 3yZqS:@WTƀz Hr[^tS-B0h]3tJtpwU>#拁4e (Њ ;n]}~1еilJH~[WCF`ݑ0aP3R(˿e6 gvYL)Y8Vh6$$EMo5CcdIlj{HZ,`bJhsI )&0Y J35&H9EPţuEf?P:q69BϚMٱ8n8bϭjKo:Շ?˝$cvR XI*ZD{, md@D\m"+="8C ʠ׹b7[߼b`pL~Y:;LPXvtY3H `z1ETNI--,<upBf}E+yRA%yGv[zy8Zj]UM2Zo:B Q%LH;Ec1^\BR&ʄ(SqW˞!ke^ھ…L ᔲ* bVzJhҐ˸#6)? zQFƗuX2'2 t`B.B av1N9;90$;dSTk>G5&p =2X 萒Zȶl^Z2dFΒb\P9EZ'ĨF4 wXΚsrw f=N{QK!"TQ KIh'A mm^'Dܜӯy+U mQ1gе\.ɤd[؃2;UBPԊO>Is̞ʢ W9@d6IK?n( 8(oow #Ctv9h#ZÆ]1 DHA yˬlQ*zI;H[v=]M=?k" l,|yG$$}Izkhp9ا3Zi H ÆcB 8XPCAZ]症tu }5W6!aȍ{8yYzn.!lRJ޵>q#e_6Hx?Ts:uI.~ShH^ibȵ.n 'e'Z^Es$D!09߄\$;GI.´el)4L"q1'qhp *VF棶4EKV:J"7{tײcΗgD6m4)@6䦠S2 hY*$44KG ?#Xf,oqV pmA'mp"+dFHe-/NJqauHq'n4;a]և7VԮrF`AF]4[q oښp7`^+g*acD1ihBP^),r:AjEvkZb$.0H6DpDGK)Q -tꃽ~Rf'BA0 J[ ΋|*޿|+IqE,,QU 7%Vq&WuydP^sw 8FUG;=GMLrVIRiObCfUXQqx:O{X;Zftd9LZ$x QV k\IswV7G'J~\~|Coؠq870'X\Bq}ܛsDDv,qD [`k9ȍpG 3WGзelQ@Mz-RU^@j)ȍnppYTL*t?]Xݫ$Z]k{;mCON2?άV ?7Wf˫89wid;ZЧҜE<{0J|N./5.Ȝ+Gozm[T1~N*o}lI# ~MðqEfBa0\ô)*:og =w՛lUGedInu6׿*8Ww//\j4,UI OY Jx8K $_IQ"czNSqT sn9?-~nCNUw;;תNWh" nYqܸpqꛥq.>j2U{qz(A%|>~|wݏ(3g;3>o h~>φ_6 Rգ˭ѽz^<뵏s5:'>hiKK0dRIZBΓl h4Gyȫtc8 `QP5)9b 9sQ؜^ɩfԸ< ym']Nu>o)ǻhbɥRJz& (."C( x6"DTpFM('V $2)@Dneʦ=@NPyi˵)I,C$ $S 22kA̼{vOC+w 8##g1eH/Y)En< ~n%d9ig"H}A:I#*EG '*xN]P1BNFi)Q+y $ZvA*# @j '7L@&UQsTN% $#ίҸ9Y5XT&-a$퉵 #<"Tp 6dldl a^@֞-ο5hk\ ܁/T*d%V?(1 Ԃ7k'ATZgL)G\Yz%04L~ٚɑdF B(:D v`R*DcSn!IԪWw#%u4qZ0W#1 B2*ZFy[ FjU!}%W+•HXE\u1t=uMv7箷:[]*sءBGFj w=3.l@^NJ4y)w$OJD;C"qB@偞m@"%3\DsS \"X @aA9H2_q6p $Ϫ&AUI !#zj1!1p?JڮD9CU-7hΌ_ 6Ni1r=3ڂƸ Z@A_;2GLuĆ7rV,XMdmS(Z01. Y$D5=O#Ly{Ks4.>&tPmō45;Y_:bv{_~iD즁HՁODE*5b&bhe&xW૕3z8kPYgח|Ux;~mz(O9Ca@7QDÜ|EɧRE0"\bЌDZHaFiVS I1D FAC =%5 i1kpĤX$ҪbZ$BBC"FFrF9(c(wjO8KNPvFN}X;*sJoƈcbw{}diL>-Ljpa>2z_AI&Ĝ@Am#c0Α LKRB>WxETFiMkΣqR95- &Bc ,F>J<8O '2J&<$:%qWki)q}`.q"NiΒ=Ș>y_nя?֒P|Ci4*{? ër}r>\ێ/Ϗ(Cc#[9u[u\b 25uyjGDZJy7*{:˵0KS~`0)χ - q GG#[UUHG &MI ^rEnw3)ě b*MjI|5qMch1WgP9~aJW񩊀j?:1bS.ӏU]&Ap&6/5՟oAyXՃ l<(z.ƬiqO!|)_?-udSOӖЛ׳\M -(L()tyņ&u*8dzM1<$|ۍN#7y8)1?¦W[<1p6fƷ \$5m頷o} 1<!C\rwl6rGFOLRrY Դa^xOg&>Lx8Č.aw[̀U_;:YLD~oߦYΙz7QwxvBnP+4ADb#$'c2&W! șG+d/{iō>޳x! th+P<^ eHQcqzDM4,dT&W\$َyg*Z⍝*T,_@{|A[ƅ{6+R+?'B{ҍ,/Y}0;Q5ś! GՑAhns7uqX "W= ubx[GOG8r-{EmI]]}WYn+}\.> 3*:WHCLQHw_F>~GlhKlw]`Y: h?5>i8m8`'*dt䘍Ԙ`3`5`UqQ8.>pJF *Z6Z#!WH*OXJuąP1ᾐo;6ync蒥p7{\ޓ|ԩ9ALgcřë01*$GZyr+\i%o9ΦZsuR4wշVdןfiq]=LfxOE}6EPmW&O tU\7㣢Qv\pΞ+ YS^J-lJ ?'#[aq=+=x^ xie2R(,KMD6 }avZ'av}dS>H%9UҜ^ p%(&IbA) Z ]b'I̳G 廟|YꠓL ./(&1{0aJv "[FQF|)xrM LŀĢ-dw>u50ZW의]-B'H}5ݺ\ M6ネF@BqYdX"N2t[[?*T EX122n&^[׭ڹ[.@G\*Y˝lk-=+1Rv68Z=Qa"Z2dhqíquBJ)UUxH4+5l殈&]egq~+rnWX耏*~*~7="[*|Rrr:6CL/Kaon L=o#_|Hqa" 0eE(IAKBI])c^ڂl=,M2L㡔E#ͤ0ơ(Pʨk^VvNܡp u`{Rv 4nmYtS?.!&NSf$Уa4-Wh.jG(n4 ӷqtnv l35KvF=M&ER:3ee$Wl` \hA.WDiEʕ6^C`$WB\#,"Z7Qjjre@K"WD{X{"J\ P\!c \hM QB(W\p3ՎNj'߉Qjjr Xi#W+*\uC+o #B`wH\\ iH>"J箆(WBP ǒ+ l _\Ʌ G䎦/dyGH\.=mGiebe2Bd^ }W55Sb9D{7yAm3RP;x\LT E]W%4A[R*] G s*r( o]1n'c061Z% Z/R12c 0P8FrRZ6rE*p+H]҉,W+Bw+劀"\oj QʐjreStE-"\ȕZrE:O8Q/BjE&/WD|ʕF?\!Sy,"\fhL]2d\ #W]-$f(r5@8{Xp"WD{Qdpr>8EWl"\`3N.誥o\M_ \᚞h]OUejsW*BT^uS'7NMV2mԎ+hhZؼMۥuݎteS4eN/ߏ^f,?O&xo1LFF,Z/_{q ŇzwKaqj,u]E#joק9wWzhNM V{E[UKxKƤ%0UF++FW]G+Lm,WO"W`,# a*EBrE.ʀ01+z6Z|tE] @֘[>sW BZbʂUkF ]G\rEy}r 䊀 "\.rEN.WDs28DVi.ZFױ >+T:*XX] -1p dhmo҇,WF̎7z'_cl^{.*@re"dzP`!v:Q#nzjv`,/JJNp&wa*ʲ !FjyZgp(E))쌀ceg- iUQjfgʁIF0H6rEVr+u\e6pJ;\]n`#WHk*3LmUL'+NrE+uBs+QjrI8(6rI+7+*u"J\ PJFrE+uT;ѪډR繫!ʕ5f6J>*E֊(r5@A9e{*6,%p*01+gpU1DO_+r;E:(`}2ڞ*6TwE "7!WꡦE:pie;Eir+3"WtKQlSawl]xF1޻;J% Z|A r1COr"\\!2Hf$Wk" S+pY(W AYHXK`#W\I>$JY(WVRtFrE7µ}/E=trEg\9% 䊀g#WdhUQ] Q`8]!pWS+r+=7/WD*( 4|ױ+ B.WD)Ug#WvGc] =<Gz(Hl `!W6CM/ش7y=ݱR \βkBv"%,)O޲&޷%ʲ("[6nQR#[تY(?m-(rz몘qkeI4l3Ul356(3\!Q\f#WD &u"J\ PP2+f4M^s+"(r5@2k)D`c$k "ZR+\ P<HjFZ.rER+Ln3ՓȕR=AǓ+gp"WHkDPDAʕ RZ# Fҝ? +/D*Ġ=#{J8ھ:Q:C@;(`P\..rEA.WHiSۅ)Փȕ9H[F7.rJLh]ʕVe$W +ul+"d(Ur5@@sJJFW9.rEV.WD)W6(9M#7|p-dhӟB r28Dr i=7g iH~(r5DC%+@͗ilv5WGUɋ꼭0YNJuq@)>NsGg`akqj _i*svgjueQA5I{ vp=fǸZ׿n&ŲgՇf^MΊ<݃"n`{^_#ir^җu?8 G޻/QysvXοs /xa?\~ o o~ /o@)cnx^9H^14/qB˫W;`} J~}q?Eo$i73[__S<:8Fessj@=z38-:jo).r΄YkLU|V)!h 5F:{UP!r}}O}P&oRJM#v%s M@r҃ ڗns%<:} iu'c70f*U4yf~>r%6'].Q^svq:'hWz+0f/ʪd__7)lL0'JT/겴C1}uc1w[{sHvFbt2{B3v#? v3:}n=L㬸yqU~ދgeqU.fm_2 槸%| ,"d~a5"ZEݙ{lr` %lmrAױݕpN-)@{yb54o A[ck9c(l {'e6Y ^h ({t(Cc'iӤͮ$X{ț-=ToݚY3]}Y[ڪc7׫"M-tX =1x;* (s;Ж83+f^*ot7|(l*aPg-bAN xA.D]yh7{@_6yvX>w^wc] t77qj1T@A3E](r/gzXOp,uP^i;;ng,F6IYB ;7nIOs`K@&}fμlHIq;qߗ"lJjn[R*VEr,mؽIAV)8Wp1UBFE*sYG.2T4KLJIs"ru&,){V{Z]5 Mӑ|pM||741짼 +0fT-C1K <{i1r(Hh9sA^)ǓGUcϐ'#._NcŤ#ĬSyi|8F+M7F\$e$AkABt䤛SMY|,yj,++t`:F Rmcab< ɳ1ӓ1 dq(C )+92;~*ǂY|NƈYWksY/TLX}xHB=9ڒջdy>,ݛ>XZ=rT"d)ȄQ\uHQ1h$L#RkHl*0(Ɂ;M%OyA!̂mF/9vagnN6$n998wVEXJ\q.Dd)\di;y yI88$8}0=́niǗ~xSjuQ0ywq!s yx!$N0JBw㔲 ֌RV0ca&"T)u<9J3d2;-Zi PZ$_?oڡ6G(3BAyЛ8eA]<}T b|b~}9 ʁA6dݶ9!LJMݶcLIYMsF9@@c<54s 9ɨ>UtsaBb*9t <@5։AC0f S1ԄYGf$˧qF#s$0 f_G`VIFRMGƈY`¬Ƭ1k3z1xz}ܳQ]Gb,(911@V1#adcaRV|K`=;8jvHld74v4+Hy|kxE6a)>({q8ɤ0$7O0gY: 9If6(f3Xݱ|/t azjV n QSx!*+)*jGP{6{r;8iaW_tx'Hܳa)pc$d-fg'>x@c{6ULLn&HʊOs4!g#/(#b,5/Uq Hf7]qLywW\ݽVLOKfힼ[fɗ7ss }qyC*A鼘'o(*mV=_+1c}_NM$}GbN̩Me$ {ʜr>_7ǫSx^ cFz!8L;UGA8jN BoG6'-\u?2yD$zΨT($Ηi#ÛG4WJ.O5% QHVؾ_>t1ד'XΗ Z!qww}0UQCֺND[/ CɝSeHp)hV_cD# ]7a`Ғo+g69:wV=Z/_%%tf+$5wtKXrGQX1 hUaOnR~)BЬV&#jw[tLw's)I/kM8*E-zG4~ߑQBh>3( G0nK47VR [x9whią"N/ruhU]g9"8aLiZM2 omłw/h5*5R@[(=!fX_m%QٮvlY|YHWwq$K܋!j .)S| N8|:q 0XDdzVRBuW6QF%:0Ѵ>MsՐsV) 5պqaHqUVv;'TAqa,u<ռE|bmF˨K։A'rˢr.DV_uCw1a}KQ\X>zE>k(ijU|EmoC_"1!G{4gZ4S'l)yvˍU|K4zUǰ;-_t k7plǟwWp`ٻ}sf}R㠴Â'dݦNtgs|Thu$vwC!MiY߬bLA{bGJ\X#X.olӪ|QLTܶ@Jt~ I.ȅC EƺjcQpVPmOw(Q V6J(< a%|_K2\Db:׷1e }*,8U)3EQbWSe\(>x;uuC=fx"ztX;*Q/SzǢL=rEbtc)br]=YB?)p b@V֐#xpriӔXWZ$ |7;pt=rhٜ'Z"Y 8Nl#8P[}X4O wz0[vP]y6FS*b/LqQ(24yo =)e pSkF4ܝOxĒabȗ9ZfgHRd.B׹QWP-i CY,*SzrYGw}6AM r+u "aAf{W cs q-dqiS[ V/xﳏK0.At\nP8F [c"y[M6]r8O,zXr[~V<"1'Tu 2[Ոq#S^1ŋ֯pgH(NE*>GRTX M)DG1>$bhQBl-cg}㽂N@Vl9~_ ofOW}M/ϻo$ٿo7wPKfFS Yb !&NBT/Ǝ% Xg' [K(*KΖ/jZKH%X1#tf" bPt KQOza@Q]R+T_#hRz()|_yIb01Yzi-9R$H]f8NN"VA k.'Gv<9pLH9ݹ߬Mh.B:XH/}ozqX\(*+PD`/RQ9O6ɪeTB>Z')OYI˴K $H[IRfk5$7eM>a$=2cE+GDi $\QNwMD>~#^Dzcdt+;$rdbci˝=۳$A C1jwO% <qlM r{R =@]aL` A)L$B}*,0 WV5݆SϺxʋ B1~rnS)dm}N/>GQ yɬ` ZTeJyEKrS~p u-{J51 >aዯ6 Sfް֛+L82k^L3-}j k'D[y^_vǾ943Au.t,=_>m 7}62<\U\ctFuNPp S%5]B\?[6@{FuN:J+l㣘@GxT= fyvU_t1.JZ5sR1~,U˒?%ՠ²e)2۾8g>5ր`Gխ; ܿyh󄞠OX{sb:`$9&s`!BkڏM#Fz.-''__^ kƺZe {7Y\{gQҩsNs7t_Y[<=ftT;0ng5q&[O+r+Uxuws>0q3' {׹߁pG!W 뙳q.'[:4u>+\)D$ !תz%]הe!DIÈ] Lh %vT4yvFK?z5q4 y^F90Bz- ux^XrЯc]u! :W'PgLukv1_5x̽d}j?"%?1%6Jp9AROoa`@ǀ1n1M'p{0`a:$>g݄7nogF4z;7cnjg?E?tYXCg%x~ͬ^%x[e|qGЫa޿"zC= QBptn+mG%}915=} >hM9Hʯh[)YTٕK!}ˊ$Urv8Ԑ݊)c{%X*9ʭ2ᾶ`{oli{J (Ɔ>-0ӟtrw^^+C?_Bx/% mLewõBuZv_0ZdមAѴuM^>EaO/ L^@Nӧ$vYau|޲'%+ǒANE00߳.Vz:-#֟ULYMOy+E>ԏ&///C%eڜgի[u2ju׿~*WEV,7 EMhJM sql0YY)@4] Aqr\~l9'PU-x9!w8Wq?>.WdTWUt̞Vlk ]lW^fBϾk=KFH!ny᮷>2! 0OiVFgPetZ>~+R*WP10OO2r*6.U59Ad|܃:q\Fib&Ńzq JF{^vQDh/ֹHp#WFv7t6ISθR&ON< m]]0 tE{]dT1ՋP _D+)EOڸkǾs8n89$FC/ AլNj$Ox$G1AU$~Ia CZi/.I"-@S<\X2 >7b}cRbUFv\(5'QC%a ]  zrf*">+hN5#Y3ʸjpP9OAY@ƆH2W8a7N*fI١3`1LxLdߤ L12bd`(e,4"l(4be>H JT 6ȳhɺ4+8B*{tKRzl^:՘I>: * FM8ӂ^(J"ЦN{ 1ypj6lmd8Z c8Oj~ +[hL8gd|%ZODP2G ҈3HA &֯UEv3+◑g0keN^ 6ԛZ[; ] 9im"=f88\A"#e(#{;SLM_Fd:ȃbyA]| ~d+-}ƨVCyUc6Ҫj 7*f)ĩ C ;ŭ/ w$GFR0-_H'Bf`Ϭ|TN1 Ho&~DHKez m(#\& bMS’~I {sGZцB!H(N"b_[P`S ˇ1>A)-4.gbJGTEUTأ5Nb9|ֽ$Ɗ ׿d׌i?fA~B9.R$0$$UZ֢(^EcDR]s6q?2Sӻ'srm.7jZWa4O4w\Pljz ;sTgNS{$䞚,ϒ_6Z?]mh2TP%8Ss%G(8ekaLh==$H?cҊNDCB9O A9/9:ZJA\hP' -4y0=wTinsu!2>Nze2gJ )@@I( q"'6L\~3 :58DAhӱsd Ͽ8!=Zپ_XA=} X!rWDHnX>ʢv v"OiG}fx? MTA@{2 (!cvMp] ԧa* yo9%dY\p!C}Ԙ5qq^=ׯ(]6%f&Ӕ4 |+or4$b[eL 㟊n\@: v#Ep_q aU`{D´Lxqwbr88k1J7f%.01i6pܯd%CGAkUۈ*^/n]YABcT`cDW;aǧh62ˬ5OpVe]<|X'Z,.v$V]j Gzf<ߔ?~E4z*S,p'5Vn.2 :G4KK<"8tC:lkHi_ZPkJm!b{ls 6>J<`YPtSY Z\f^2CMI,]Bo)c{90?,62MjQc)!,8N (B\3Mv r>`[ƨbt*5>Œ ;|S~ Ik Rxh3턎8j_zYbW >Y5Aibb_ O~s:xX h0PXd, l/KsZLI 5'( Wo[i`;i3a`hqlz]۾z4f>G:B(2+nVl8%cj^WŅQso"5xyOzy UÁ> "KQ,sx-4.Ec"yjn9.+ZBu, 'o;ѿ[#k*/VU\3Ъ|} Ôv:VE& =νƛ2 n$%ecL@ ܦPV}ș A<ĖIAs)?6}D lTP껠b{0Ćaҷwjy;ef r0-=)m?6VZpBZգB'̣ R{ohANe7#3z+x0^OCwL$BJ1 oҗUD6I)zQ?&&:2ϪB4Z&X~Uz,0n]-wEqU(ɮR)b_u+˳W2όZ߬3<+V SIX-!Hv1 >%QwϔPܾ$?_ (;.7\=ta[b%fi-?Lʬ2[o%qc\0u0u%w7m~RI2T;0%;܀r]3w>mj0,Rj./(s:~ ͸zNE԰XC2j{0O@ .\`u?ߛ}9lH搗&,1d.(Ph*w^i3\Ϡ\F.ʌWm8g1OSEi3  Y.s'~U٥9U?ofKd*у'k()0"auvǻrB;ư8/y̡V1l;/>ΛUok.3$X/LbߛҋuD}bvmRi [8)QY9 =^"\]zTT]1y=N.#nh@ߢtycʪu:J"6O9` 7'v650D`YS b^.+ i(YkT# yfy8_+C(D8ykF\ N(zgQ >6IE~S-8hISM4Iki-kUQLR yt%ɘ0ZQW,뵓. 3)]b`[P 6Hڎ,fwe>ҘPPĉI)8HXѬGk汼~DbVo5L\n`oW)ʐkZq*Pz+ >] yUi笝#]BiT3t;B 03X,wX/FLcKF;<' g&uH#'RDH0 &6ڊFOHpE!\lV |@,/H"9OőJ&& ,8ut8ڐlfnx'DAAv`pS5Ҭ`s % ߺ7+*=!i=6@ju-X ݇_3Cg瘂,  ^M4##3zcZƙ}`-sJ<'oV8P-7^\(T| w/Bq98"QD1;HldAR)q.]pAګ,VtU1p8ꊕ=ڹnFH4'0Z0m<hEs"?`X.hqjYjīCHuZuDdﳊ篪VCDV^ÂW#j큵 bq<c澲B}5F tYvQ]#" [dU1K.lǎML@RXLd_͢Z.Ƅ.ʬ/OCi-4.I!Z[M~Gcma/^,%§PR'%+ſ؍`+.C6  p(Vhǃ=cJNegn*<1`莩^!qJg;-t0' " +FL1%G¡!TN }T|)S%sۑS2 ~LjEOGP"p(fOY{0Ώ; u) {=f˲Ys Ave/kp}j{y)Sk O>^oVTEu^~՞xP;Rߡm[%=`ֽ: &:{twRɭ =$I n&$tZ̍NZF3s;F(c.hW/6v.rV`"2"2^_;z]0⾞:w]`1$'?oީsyY pv(k!%*Ef + fRU_gB06jnO*śap9#bjbl~sQ~e"4_z+5K:pDaX۷ɹ v1kƿ&KdIیj(Ò6㾧yΰ u(X묿%40znI[u`4P{Qsǽt}ؐ >cH3XԑW97gEQ돀 ~ Z`0|}FwY>3 Zy0C+S.|8cAW02ofI-'u:,v.\r,nztaۮ? &FAԂPwɫ`ikz@'_"|RO$dHr=a}TBd}$7T1״~,TXH=ίR#$x SS֝y:K;਒ZƱ\N\V$qrq,![/9ըFsy5 Z1Sp G@_Z(a;ѡ#x.1b! x6w`4x`ԃO& ZJЍܧz̪̽:nWcUZwhǼca3c(øKG 0?]%__hiw=zU|TB\Q*xNced|M0,7;}> P2ۘW|qUU3k̯bzV_&@IS-wN&?/W8o=USN$,q|GvKM>d7쫼_Y#21}։G:&Sd6JbiOWzk'R 'C0\t~wW/[mwivgn͂30#RP#nFшa't5xn>/$Pxum`e➭ "r6LglAnn 6Lu]j|Y"jU-Zzή>>IG^5Tpш]>a˨e 7kAaDQ=:/hյSv: c 16RnkJ$xm0,z;(©7vj mϟ1lK6_W\EH8AsaրN%"I@VR ncDZ<#$TɤS T2,/Zz9 8nSQ'24wz?}_8,n>Up]OYÑgdjM.9g8c][@y# V~$yjHtxՑwG~ip* *n˺a0=1E ~.!HbDv6P)%4e a\]t0{SBxx09?O#m`lL^}ϮlX<+"|N@/ 7C)r^ԇ~94P-3hB>pAfU{)A#,ɡ{ sN U@,td@{Vh50g~~a , 0Yf Eptc _J^&)WO??2-ܣQ:  fzas`ヤj`dNi%B2Aq-CwL֤ M)/Ļt&!}v:[K */zV|U#6'x"RI?_]̹Yl|Gg,!#~1BV5qN#7T`e?I2!!N~˫ee7!-A;0܄A6؞)aR+Z@kEF!V\EgUFwNQlnוn$:>;!1)ZGB @Ptn4ט<70Wssh!$䲭urUM{)W1Ll9]a3FI )jWCˏ 8CGԗAh7WRr d!Qj<ڃ#UP75ul9 Hc ϕ%|_ovxGX>c<[ў+5BkP}U7 cnacܟubB>8! -RJ"KQ^_ `̐ gN0%{%+dy]b &uo2"nlqw?,slf΂gYyVY? ADu2BŘ<%CuDYJ> "!Hf+Qc@w \K8PJG'Y˹I;~ƅgv-co_}Ɋ}+*V<(/r{ 3 _:ޛ*^NpfØ/[Zίյ_5۶iz֒0_J!1&!4;T9mSt{7]*?x֥q9vLK(ٿiW p$Íڢ~ѓ JlSIEt%qƆܮaRK_V?P1U6Uq~k⟂ŲVC~Ŀ,Y/l0^NS+ )>Gq :ƅf>\=+)gJYrկ= $P,<(q$ã*4ƹH$߉Gll9*K. )l0 9wjZ5+P{ <6nU# eX"j̽XdbVsn$d`k䑑{d2>'+p>_ny 0"UJd-|ꇕ.nysSʻ1M -zq>̻,Tɡˣ%=󸂡i_Vƃ <.#M)F 06n&N&Dᓾ iD5FS5\ݪmG>`,-m՝8H1j7˴P/X#r5_tCQ桧) 1Z/Ŕum60+vr\(hM-`e#/+V2 wܧܱ0AL6A4\H`Ԛ<5~Z$6 -YEm4ǯ.`TGȏ:4\R^ \@"ڊj6i9"p]}NX{Tp"-j7}|jmZеi!I(%1g6h)6v T:$|Y N4 NRb::JM}H.2 '8E9\pMT 5/P}~Z*geHM^cqjE.K$-%E?.!KO9fcp6&i0p)4@P o% Z2)Zc|ڮXWxOy*J9X  IJ}$E 0IbBHDܾ{!F\)d)1F 7k$PJK/4SmT$Fciچ(gN 8>&T6$58DуSKE/SlY^)1=goeeK)Z;{3NM> @eΪ3!ό!}HUڶOSUMp,mXʵcfF;KH Z;p+⼗ bby VsGr$JT+Páo¹gk'F!X[= , .q+"͐K)F" _D _]^^7RoS?Gi~SUNѮkǦM0(D3!̓|D& |`ƶTqTR^O_ۻ֦F% >"-iŶҞw/xϳZ]f ?֎o/y,TDmi&*ʾ-K{ݎ 0)Tޮ{x_-lօ.^^W_P.ڱnw7="v{jO>i#z/jr}sH \xȝXQ ⨱SVʲtuN~7]T}mLkm#GE˙18=ؙ>dtز#)vSduiw^b)W'_@(Ţer{!][ VJ?B_)RAu{n w_,ꏵ Vp$͙ W?jz=U*OSEo_Nif0o)*7OL~gBə?_32_.~|.z?D߆ԏdX.7\~l*䬉 cNtbkx~HMY<ǘʐ,oӛ0q:OLW_5X8MUk xU.Aܛ_t`{ӔtVzOy\L¸I N,XD2[`kSYURɗjD Т>Ilk8REac4 aXQ;, kCX$* %ɂêĖJ{DXVjXkXhdƨW*\!b[LtmVF¯Ff-L^YB Sۓ"n9gRp*eБPÁCd^YG 0;la3ڂ{ElQ0P+i;F LX+ d@mr6;=m%P$ %GX!AY:uMM9kiIX9J4tqb zQ hSc:T' 8aJVgGI RF4_JqJD` %"tT&BjLLRgb&Hpa}1YiJS!\1;"I V+:SxL 9KX/T ڇ7o4 ]xj˺Sqh v9Ah8'YI-(XZ\$:=LGFIPY% 0Q2]+(We @vwwdۓ|yH{XwYb)@(-^gJ}fk .|`=,^m`:戂$=67T=FT/@:iRiE !&#H Q9/@DžXü)!dH!:#UpPD]Ѥr !HG)!t (J P9v+'mDCx1Drek+a`ong+G _ym˲: x.g:JZ>a@ `htB1bi. -T^B.>/9g+y'u6:$3Fq3hqFJiTvC1v<+QFj7MZb܉M72@:}cRGο˽sD49'9S(N-Npi 2L@;c:JI|ǯ Gڷz<"t!8,n/gA#elвc4=ގ#p 0vƢO)_4Ix{`})aYf*NnZ> sGpB\# * %Ȑq@I*<g)S8, 6Q m@KC9̐DhdYuP?8^(d09GZG r 6Jq6!Cp7qiy;#rc?,'V>n2wp=nd1="6BލÍ!< *8:hR+!RS&3a)Dn@PC7W[L<:\l3_FbF޲0'`1'->Nx [#a-ʏ VX(A0o= *[XMִCH@JhtsqA6Gt5mӾO}nk8~pˊKy?5N| ߙ_&0m%½w{(-t2JWHM^8UE̳87S&? f*d >Zђ35s.+d _ޣӳtv>/.u6Ǥ?bd=3PE!dj,(gGu'Ղ[$i)n2<ExX62Y5v؃T LGd9lg#Y5a-y631!5PڞcB#K Wp/PHHJ*dܻ#wڄˆB)9rRt3kVz}9XC1]~6 +ӭwG‘F}b 4]Ƌe)5Әڿg~͆:"喃"~G!1cU 1y~϶֏ұ,m n{uUֹ=9 ^Y5VXM0>6WɽSJ"5>^^,XuOa\H1O3qSP3렫1{ߘꈕ3gN{ ~!%ѸG+V>!ϝU%(454j*K1?beu:KJHWdhqՇ0k))bF>ԀKj΀> O%ek@a [c\`YХ\$![ؙ͸'4gӢN'ۃ M,kP@A0W e,8y qvƈy6܊UHs*9[,o%NdU'Q ^\SI?CP>: ҿԪwF;Sxń0fb>B%b,lak0v .kB uœx<]wOz紩ɯ+ <57OJw\\~z;]L}cO_ 3[;+[ %2n*ir\(h^X4zYjd{v? :A79w~xGe6gF~. V!G)z\hc2ݠ/,[dAjyi:2| Rj`rȰ*d kߚlߦtwi_!_ bĬ" 퓧(P.b2^`G1e6|:/.u+ן..6t#97 27{-f5}j 'P̉8 OYNfrZ?)\-_̊o/&ϩNbԋo|Rnrhe|t);Mq_ `Y9iFfų`Y"N2$]0zBVӠIK}pr#&&!x#+ jh(KVFAWEpZvU~&~54Uڑ_}9Nxqud $NlvdՄpQ>` f fLWj˱Z~mK;juy-7c8.Px&P]/>Mk dux]瓍:sʽ֟&.Y1k26cv f]} ;Y⟛I0DI0U.[p/,;ekl1]%n%|~}DO!&1XcՓnMJ)BƐjr(Ϳ_|N./g>N[ GxпBKߟٛw":'2ZUxaɒɌG]#BȂ0Wxyf]mwyMƱu9%]RC!/xWIK?䇳;cOPs):yݛAGB瑄*"Ol"1y^T֭c.5[GI:ÈλEGC8%/KO#aYv~>v)u_׋Lr+d|g5=}] i1|AgkTcTb]mvw —өWjA4$+`cK\0m&7ˣ,dWWwN̲C3HgOhPEvAo[ro/j<f^bеS#xV+UUȸT  -M#MO*E7Vg9'Ĭ|30_V#?S_ɕr-rO`w^RwT7uϡ )O*QzN. зIĽ=(l8coRB [s{C{׍t_8/{3Kk}7cZgØg>5aW0W]Dkf1};bfӉ_}@^&rM7Vt[N8Y]7m3FҖ!eVv?iL8[6AT Zmoۓm9nkx$Mi3-Je/-mV#P'PW<woad5})8&T҆?6 eZ6>囐c?6sgk6ƴoy)mG c'Y9 xPBi;8z?p4o{{=%s^َGlr{%ȉ!Dccۣ6xGIuۦgz"9_irJ[ȌDaZ+YZckZ(0Y?ť*%2"3k}g@+Y4b4D}kѳ"ROA?ox2eCj+ح8`gl1$8?8 ;:?+w磯>e7+JdwyU 4{vk·;RRĥRD|/DQ(M F[<1m yƠ+Chorkoڷ_k;]3ƻu"%AƳ qkTsf&}[ ׌HFm^|~^GflidZw"ė\LлMFMvKq oֆ}4 XN0͹.ͰU[+vNam~(k]Q+Mo(yQا4xLY\#|o#I̎> `1E$34W'zs[ˬ5 KTo1< >X[򗍿hwڌ>hx:;5X}4rیf,f.*71cޱu};48> 65c}fZɡޅt79fE1R.\ءc"OO78zMʚ1n{2hϩV)+kx+Zn4='@T8F!4s'џ~QħKJ*]O\QR JEx@-Fj[q׶8IlO[4R^c4AQfj6l0{%d}΀F@4/cCl4Tڽ&&/%hMЍc 5JfJ֫_l aBÕ08hx*L``EǰAVATs"J?+"ThgӦĒ׌q;E`hFbڌy득CwmsBp̣|yFm=QJGh!<1>g,5$O ] Vc֌&Ƞ'O&ɟ-  wF$!ӅVesWe@{GN ˴sǕm72ZGMv/pi*,37؁LU14$8,TMR (L1a|'F>7 NR:ve6+UVbID(&_<L1b_yv٤^,Sٞ&S8Pѯ> q˸_`>rG@QpP]= LO+ţN6ƅd"ƃ\"!#}{ hMR(_e6k>^I>(ywXQׁv'n~>(@: 'lcs10vKZY0I2_gfN9>D-y"FMD˩m9ScqtfR=9EfY]k'|%-#h+ xtBtSľ&#MGX$99/V4$rpg!$"}Zuq]^, *ē%m`h17g#Җ7"x) BnGuf6Ti3U;<NA ptRTki](I|Z8ɯH&6!ߓi YUDθs߆(m辶r_OF0BlIs>u+S4:Q!֛#gi'>>uqV *C6`h)BW>|Boks}wlύ[fc+.rIJ|Fwhr$I$Ngc{[tW cwlÓA^QrWy;vwIa+ɼڇ϶\q3҂+h0mk;LlbJ?>a a(IlͶ~!k~@n\>GM$:a򚶀'G2(cq%BY0vJIZ{uk?ZA p{.#]>t?[墕XVV*+Qym%)@ۭnEIwwM+DW>+ h=$ 郠kS 0\ yRx"[Y7HMZUjU U%V5jmh5Չ` Yw#J Q0W?M}˛V%}4Vr3._rچ orfq<ID &I d2Q C ݥbFcֹ4e4 FwطRr/~bYi{cq ԰莱G+G/OI@uX]ү.G0iNj7`hFLQ|Њ ˘,ʔ: IiᶬkKC-eYN?\~p_6g"bͻ| .xHJr.2Qik-ZN( ʆ1I),1`<#&$^*)kfHV$@ͶYl 1,yhӯ/V >61ZgbvhG; 氘V`7Y/Tb:X2$|s2dBbd-G*x8I !Mң裆iZX,@))ljz|6'wU$Vh[a+,6)2V!JAJGT[\p,"[=h*Wmi'7G_Ldȉv.`y`q `)E!E%GO{KzEQjӫe^pfO`;Wӹ%ʠCR},eMB4N %iqV/t- S$y0 }dx@hKz%nіvOٚm"#VV~ޕ&YWuL*1S16*!sAe\f Knd)RrZBoH+(%""D5oA1HL&LsIz)0E&*◂eTJgu @.6ڲt;'r%f$L:9XY;%ZpH1(Ht(VR gUx#&u˄v7FN1AL?+?yGNtzLu)9JHf3qDV \q;LTwP9.^grzyFYdJVv ]{DgC%Zs2p[)^\y*B.EJ](#hd++FNed-TL).2ɠ}JgZ׃*= Lv)Y)UtѣX@~jsi.. YeƫgDOyY9&_!:^\&Vm\:2ZQk&]ڈX6!(J!HAۤ =v(ֆ0 Z ٹRYU3N.{6wla3h"yVbڨ鰘J.Q!/Bw:vvVN[v'?_/z**?|0:f;;m n6׻8EOl|S=Ia~:)JX\^'7NoYzz@jS岷bcvv_K[pCӯ^4;?&.\qeҁȡ\M4Z7ʧ6g$@FG 46I. J V$R1kQNHRF{EjimOFzA;߲6+e;WX4.Eq k6 KK&SWVKRXB1$Ws-M1o=(rTR:<@R'6rz\tuJ_UG8Y<"CڈD(t3SSB`++m U.@=L NOOsFqȂfd= 3}O-odH8^u܏ LnZ2 gTc;hݲ#(g,suwҺu6V7-ﱳ!!_g;&`l8£֥F[kwP'ǟP6+$6p>z܍SDn+>+⌍|o(IZhFkI-*$rpiat0b[oQDwca^FJX Tv ']!HmKMBsIl# |xit_u>]Lg / }Kj 8lj]_AI ck7@h0ߓ\qY63r= r賬٢cc$ NY>99>BoˑIG[nny3F"22a Bx)UXIJ墒 ؃5ģI$-b+GpIwQwu'7Ș&;2 7o\g/]l#< ^^|zTAY4d7!K.`鲸*VX ty3)tLCq04l}H-E8F^ 0ho[e ̿s[6Je:8y{-oFʔaJ!>D!z<$X ߓFӌr=ǑE/yHZ&H$j^I/pk4#G:[-4A>qe/=y{1fْ o.w0t~)T*v'R:v=`b'.5wh.pPMc;`ڇ+xop7Ndీu?;1$`ӉS~qN #WGZU[Bv>IO>Mryb)4|2#k5u4{׏o0fKu@}#\NlMW@綝Pʍ7h@<xCQz<6>Iضy^0 ~@Cܜqa}tJCnN}2aFIdq)yd+]ϔ;TRA4v[ ? bZMIV}=d+ѿjqE9ju@T!;\RHJq5(ݢ {X{FCNջ!'ԕO2DHFbۤ"A~F˥ږxIN@䣪c `3ybsD 9o)'|!TcJn~ħ!gzkd>`M1(MA؄ MN)F 14J]3*5 K*ɠ06>{ƭ&>ϦnP$ƞs8K!ֱl)Y40QYePQMq[mfPxUUem9V*ԃbuZ8bz;Q$1SIjYXpRAi(Dtl(X17(#(b@;G|"/|.-"p! K]MutPT+Znb0G^_t/mUލW t2Q !wi+Aͮ :i1)IUէE<5YlK׻.${يÒ 9 # Ttg[?j=}0֬Ô2_rqK9e]/z2Op'9_^K3"h$ܪ2Ȱ;Wc atoݽ GQy 8?{ ] ~8،WH#u 0*/fviЇLڨUF-'nYCFxJ1EROz[źv|z&Ax8#){1=bn/y=IudN#T( et. CgPj-% s{-l-K#UP2FjW/DJQ }g[^|y]s}w༽0jmd/Hs6m؆ gH]m,67g(|\NJX@L@='F{]Эx@K^ tcp䤗3ٻǽ D wpir}$Hh_ί$z잹@U6_zzD/W^^z~)fc' կ8poةJ$oK7rċC0KJ&}q*"r[ AdI XRF[ϑָgO%b UbfH! O1MtDBQ)ZVnpl )VoE5n9Ls`9?6myXgc(RQXͥk F]z3gGQ3jdkM-/hBkκșmYB%zgi)%eLGSf~d׃.jx)7=WS͞ĭgz˽{jHꄕyγdMT7=рdQ*>(WEԞ 9W11羋UVlZTd1U sugm\FӇc\ki~&6P͠]iS xa\X[p[/G]^_ZT30=2njӺi38c;&~@ހ ^A˟On<5kO 0u{r vbVXu588UrrHtZw1KP}gKU mB`˜T*B[_(( "f7U:?2'N!7υ;E'G^iSYA9o.W߹¬_凣e?}m+lM~[[ښD6lښĉ _hzxZܾg%b\XN=o't0%nзJox۷X.q77Ǿ:2䃗Ze5fz-gQi5^yg%$S`NqP+&=qM(rNԟ@-ؗ]tSOэ+}`φxլ*PdBfR\@0NоN#yƢr,`^vOX[\#0JG^N!*I0gb=%MF hgA_I>Kkq|Q]8?"G]j$3{= {{ASM t(cU7"nIPR1f4DU%DF~[[ja';fĢBsXxFcA/ :Y}dqd#QUo"o Ы ܰ^MnR7o$ʵ~hTU1ž}RJo~sPe^Wn?ݖi؀θYͳt@@ZC'76l;yp'IBlǾ_\MgΝ.y%!}69F1co ifZ=!HK5FOE@sؗh&6A:cK4ӚĐBF-bd2(XɑԼccYF5%cu,{| YS+x1XHÕنlSJULz#%! <,1R&UoǘuS4wJZ'RsUnMl+jM gV′-gmR%r!6\rB[O8gT"QRlT-Ɇ%dX,V_IrQcoE2cя9bC#(;'2`%å -%7VbZhDlSܪ5x%v"QBTX:=. /$*\L ZVb4T1ꊮ|.ީ3uUwS%kd,:؇lV t+'5kկf 5٨ T6Tb6x_t|$* *TJ ͹ءu,Y r`Yy8?Sȕ *rH{d@tm)N]CəĒ]N=Dɐ،ؠx4$BUY1'[V5)0pv"&&]}m 656[ҍۛ|ukҔ[P2M[_-Uys{5ʣS04䖕5b*Y-A,ڸ9Z&t H`)4sJ^w']ElZ5IF͂)B}P%:["TC,4TOсWby:{]0g=lT;bnRcrHN>VCJ}Q\;Ӻ%lj͝Mք6%Vg[ xH)8Fy@,Y}뱖(6ZRUBzTG'Bp,+LQȯEL%r *V*T\vM?94ݕuM{1R8'Gh4!6KbI}% /bck̦ 6KH,os2uԆIx}tnSnG"E=шt;;8rɞ ?$Ogʷ6yh1oIF0_C~8/xfA$Ǡ` =JckKW3ht㎑ kgԇ +{bTr%B4R!8-,T[L5 lLYnCfj8="ۜ297>_szڣ_Dʉ#?.WFDwWH= 5AtⴢUu3)1eP }K1QAj+-uZZ |2>4+}  V#V;՟YC'G8+>Avb/g=-e'3^;]ξܿwsq ޝU$;QjifݣE~Mwf=Xm ʏQX T.^PIft՟zk~/LG0z<aFڷq7Φ~F׍GH(g.#wC^ F#G`Qrk9rf|x6/~_o{,xK~+o wg`q8n~jfa /-t-Xھӧ^$'VSI߅myʍFIx&+$_~9dgK[~N%.MNI5[brg1[c|.Ͽ]䪰7Ƞ.MՅ<"id {WKP0exy~ZժxWOMeT}[Ѩj)ȚUIJGЦ 6SwZۑ#&Y},`oBñbY.λ.%~8 `a\\7wuha1XGn\,G&n6 Ӟ 5\$׫Zv\ԋha]b]6#¶˶>k|eK]Ƴm.dzw=rV[f_cwwb|&;wA0`@AMZY5dR&wEO٣BsĒ94Sb*4] Q.Uӱb Ek7:x"YSҤ0Sj0n4Ej,hBՠ9w3ksʼMXkp3c: X̶a R_iMm E(ysRUŲ]RA+e]MkVHC9Vd}Z/X7J%2]Z4z n3:rqeBWW_WҠAW!YZDBe=G")ʋ4#G4F,QX Nɚ  6g1[S>\୧ PȻ58dْb([?98GodsFN`>`gʐJ*FY՘͒1}ND\KG+7bl znHBlS$kR6l9:ʵaQDEdu*@rPr6Efd򱚄-߇P)'@3|* z&Df6y\ĦL:fzp}AgLAӘS;/VTL՚6_SR`U.IvmY4܇~ zd.bٻFndd(0ٷNM t(#dbGj-dKI:X*mouF䌛X) @+4c4i,f+g=$|ǢM^k.ݥV[EcfR*%{׀7*بXgi$N)z$M?6jRLѕ..xу0yf^9ֈ(T-G^MDDI&)~i^ V*֍L\3j)&d~k[%U8<7bӺ+4HAag>?f@.wT7.dMU>#ǧdܘh$pOc30P=|5N%nج"Nẘ|j{'-+jUS-JVm{ᖍ˝҄m;4Nv/Wm4vTW7.my>?꣦Ϫt5}OZ?[󻫃ׯ(ׄDuc!u?No[6VDh>%\UJtUi*WfZb] -e++IKJ F15\[߷M:ڒuN}Qݳ"AjOvtB6W~g~w/;L|ozK]O٧py1M=:\~ oN:zXTkd݈AV影aӳ$h3s;Wcq8qGF@KTI%#q7Ņ'?xTbzqrV^>$[UEbC nOa`zvWLA镠oD ˘YRjP"{51WDC*7BdN =,Z 5o3,QWUz1JeH~>;Dξм;Bޑ."BVYVdQ *ѓR{CF㼴}XKYK:?z룪N*[|U1HT%/٢"cAV}\]xkhwW̐}4;JiSTNI%شw!y V[1 $V:bnd2%m TZ`E/$%-%J+Q!cԔsBISJUpcE$}ӥp"i4Fuͩ/6{fH.:t? +$Lރl?YfH7 kW& lYv"qv~&.Glz^[KھO *^~2&]P"F;-'0k< wCS\#(3[{( a\\~Ĕi滨ԅ$NoOGE 89+u+؜ av^kY]Ncbv+uPŅcwn;o@y],hsHieNì-9鲉p?!*?uԊwq)Ɓ#7U_0c˄Kͥ\_min3]ʽ4SE =hhsƷ'5>LU-1#o!9Z 5͆Tv п%GLeR癹nz<;GX8^oVƗz2!J2٣c# vDZ낗 X*BU/ȟꂗJZүS ǭ)uٓm^wZMV㕈COvZ'ID;; ^ ^cp]Q}L/E+.ΩThx#,xcZ*XK|C< ֹۂZ.uXVV`֚o H849c0؃:Q% {?v eZ}`1}0f ˝2ZG&Ԫ7z] =< Szn h*AGqתFc[tyFu$"cI8Xpr@*>-bJO寺;FV wCqi~LM0ͬN@\&y /:/ٶ^ۺÍ-^xCT["jv"~wӗ"~V} f-÷gWä=\\4-nK%-*_*9P1I0c IA21]|A qEnr!2R鴹Ln^~Ď]1vSw=ftB{~zmFr:)ݺ%,F o8$*78,$gvGlzqA*.phMr|./7&|cB%O3s[=}M%\5%/r2x~2pEY%Tp9|4Hhԇf|;P᪴ pt쨱ڢEO+әo֧ Ы"" #$+l ^2 \2|D":n|W^O×7h=4B0U\;#W ddV$ Z-Nߧ^|:ǰ ggFEYn%&,^gz^sf{C뻝ʻ?K.%.r SvLFp.%$ h|@ENGU1UȾllE8gT3ULd`H5$e(Ltab0_1J@)/gB5ye+/N0XB9;Jwt/&*tq.b:QZnb,C H*Be5'(1p# *FAT%FKBĔ|1|qfYwy=1x_72&^zp3L;;5W>̓q^Nt' *Ǡx7|W %8/%tJٻ!co_L79f?39v-A ޒpPx7|\~`SJBa{[Nۣt5 BH1[ lI^UA`Ed6YxQ=A)܃cH.ɅkόojU2r qKm9Ԛb8DaX-IQp51^Ȅŝxbo`5x-=Px4n=WqcW[\^ymz-ĵ'$T-Աq|'*F|ע"TVӰP7Nq*ԍSn,]?&E!r7FPI(ƕ2VD/E hAf+?&j>"xvqL]}n Դ_t⊢m?HtY߽gky]Z[Zv);Nq펛ݦZ[tIAE2'Î,sC+m.]%%ЇLg^jr R{9ב_h:o-r:UG_b8nxqkӰ&ݤ)u0NqJ&O)!*((ᴨ#Zb+ZQs"$I7+.? SHR'zҺ6)JNb_ltxuqqՒĒsb F䱴 I圉¢6] ଛlN5zv )gfQE {޶-]lk~ȇ Iro݋Eo` gYr$IngHJd,.DǜǜجM%ޭѓ }A,WG>A Nevj;7.N1WR"P@5P2Z=8 g1)x'T":=$ő SʦԞw˧ϷuʨdA cX ykGMPg ~h|!HP<>ZK=[Ӵ0p >K>R,^sW'#0Irx"C*S;9 UDqk V V%Q(պiDIy(HP91kKz^°p]?bS+,=yY1wV,n%Q}F6쌚ad6hQЈ[Z l >b*P[QQׇ(loKn<1:<!./<!jC/܃9~Pfy6q4kfI]uwov駰@ȓU\"|8&\㚑2*?t3Ytp"u'f{Kưb%h?"lg"֎Z+VF`ܮ? o))K#&j2w&Pvi{.$-M4IpĈےb7o4ӂS:j ql 2&:{@[5u(d?P_PDʐh7`*^ \0&5duj80QB ;wHHo Gi M5YJ5GJwBJw&@ keTDH"Fd jÛjV9JTU#D҅eH,>!0D{aLwW΄w:B33GA$#$+׹iyIl bHt_"GnE)FfQ(=ɔ{KCI?u@ ;ʦ~T̹sZsB8_1')K Fp%sm_$r$6n7bVLU@$dG1.vՠI49"qqR 4Wg1$+Ϟ{+&n6P œ<48p0FNא!M9.w&_w0 "|eAc[0%`K qH{a^J 4hGȣZ(m.ƨVx(J/PcJ-C5(0Ũ&eǘ܀b)r/;ȸ7i#uepS_=g%4χ"^ΆWGF#G.@a#"DKxџ8s~1#07Ϝci3?j]HNb0c+RfZW˚ ڬ@CI˺u."Q(N^ 5B0, e6a*D\~X zꗲt}DTuT=Y:Sւi4j]U:;r5۰|l5Ю:*xDZ&vDmEXŠZ?̉ۨ0%In89绯Fg-=Z(ĕ̢.ҍk{fA2=H6 =7I,6~\Iy!L)BDIw>rLBN1kK?$~?ǰO*^K+7t=95 (!m,vZpXӵ{G-ǰ tv9Jg+xyJ>~\/fT~d*b4չKlX5&qZE {f0D"'ĢK.N/$?W(\fh}mq~b]}}LYc:@hR7X,"Lx\V"!jC:wNj'UMύ [~ W95WuXJ{-7aʱYГctM_^Jܾb`ᥳIaiP9" _{LI>&lTT8:/GW"7{`/Ww' C&W`ENŬIݘdG;ژm껙\^\z=+??K7goIir|;gHӟ-PV.|{aLZ2Jo,'/__~Po@_xWuxUǤz|'`32d*W9sj0.Sr1?`0<ի5?ܕ?*9io4< omQSr4^~(Rp h_ Dï | !]rcqF؊еdefMX~4v8,(,YziEG3^H#YHXȺ؍~`eݪr%}ԐYAbsq{x]߾wog:zݜ7e|MKksrldjqsxˍ/f)EnQL1sC8zm+T>|]Zy vyE廅 Ie^ ߊM xb|qLųNyD"]0T=K)XG/= N|o_4 +i;0gtHݳڬ,܊3OVk b|*vAI#Vb*%\8=- g&2O cd2=hKd9e&HsZVKwNՃdS ee>˿ZɻZo-V^xۖPKs_]3ş^knTl6eOwFtǏo~].t['lQXc 1L< Dvp V#L^69[sKelЩ:_iA8-g}x/Ӌx/EeBd\H$B>J `9xf'5W*IZ]޵y|~@'s#={+7ΦM-G˖#eˑr9Grue5b F'ͶxJdMlS,t0ɔ`O֡@c_4ؗ3vTf zb:[rΚ&d"U )*Z2aՐ.(ȹxPYUQ:ZNJ^;ԡebASe=04`#s]cS^J2.K6!,?5cIa8t(PP"+ZQtƫWgh*Ql29J> k#PT#H),ZD#6G⏯jh4Gg˺`G c(rCVjb*imud(tF*IAPT\I#ZTP)&Dm:%,W5R+XC:و%*4_֦}"zOU. `ŊZ1)U iS9+!HV1`Ъ\(yc_n4#XD FLuVn7s.#ЊBobZQA9~``ѷKM5>" 1r{FY[WFr}a~{031fs+o&1?m~RWSЃiX0 X \gn]0_Blj >{Jha%6$ I-Y5^`tb#(ؠ3Qeq '>3GB'K\K70ZZ4&}'I2eF^.~F(OqQu dĎ>HsEcQaM̾X:G*):i}El=A lda皟+)PZ=j]7WKdV0z JR?fq ѾH~ŠVcݔp@Py)ߘR^MJڗ6rx10S^GS%|ͻݜuo }T>%ͅ,#sl?߾]}O~o~a~y5o?w-qA-?w-w!bW[TBs2OIb jy/S ot˻W3%cAt{(| 剛Cor7J}znQmE?}Gzv N:i5_^7!خ+y؋"8~hn$(\ݝ!\ML"xĠUYZ70j!mP/]%66.' ݭ٩znncvAG)P:v[?ذ^{UwmPŚ.cc vvN%t|dK;]ignjmݝ%TA,Bcj#Z&N< WMƌw>&;ю[.,~>u_;b)7oh8tCO|u=תţ:9V옾{2qsz*Ӿ<)XC h3vXs3h`XywJ l^ٯ>1]u:o.5'[Ys WKXY5׻' 9L5"3דw<>:|WfuۉZe0͙B|FN aX[(/#kJcިjF'yPոNZdǣU T1s7 Fp;ň#ADX'߿/"MfGmWnTB} ۡ>Z?g=|[v/{ehtyW`{bqw/XX]zxW?tkvďZ&-`bo﮴{/>7wc/ЇIԵC¨:3EbⶒrqW-/:Yl=dccc hkcw51{c~0ICI2KA8zwP@J)q wG Yotu=|WG v7Kqs7psmS')l (:1<9i 8޻=jpPpN;`TO>y ^Wdk '梲(jbG֪`6eU.0=2ݠx{1ɋ76-ԯpw{^"_XKպL 8aaM,+vZ{蝉>`rF OAAŃxb59:Vڡ>xbAS#zLKŠ^gJŀpLq  v(V"!h$B֫^Բ |?8jŒXNM 7TJs麖TB<&ޠTL1+*] V#{DMC003V045gO_d9\-ÐkŶ&f⣗b6FgWr N& ' }TBQcN5ɜq<=FI Jb72dU *]DUQjmpIpm Wd*bXIc̃ pݻ7o>U?¤. F9F^4 ĐKaIP':Tx~MJJSs$\\mA"h-ժb-l@]B#PJ!F(j-.h4Wi;1C{Qn^oƪ4ǧgshmZ4~p1H|H޹hN&rEb)9`&dH) e*vhr/SQ{{֦X>AʹzQIV75E9-d[S@(V*w;4VoB@b%`Y]/k=jʝt$ T5PeZ.[ء~9=UA+0 u-5l>۝wb‘Nɻm;c=rd{xYXӼxyIyh,vg-6lq.0 YZ1ݙ*d9/s,M 'Y ɊǦ$rXtjK! &XBM[Z%DC߸ĊNW C!ʣU\1:qܱt6fnQٻ涑#WXK |U[Suv_j dƒ#uE$A 0˶Ht4sۿ ץ#)?;?z8SL_RsMwRbZ . k@.ǛƿED>8Ak*={Fiۺg]!ah) fS$1sOv8:2dM]6ɍ=$_fpUG<::֕ky liq9rx}uD_tjţ$2{l_q.|p[WieJfEo|*{#נ\-ך5h*Vc&t`D(%&F}ʔ}BƒxFy T)1QnΊUO ҃!\KQ~(Q)a)*4Qn4UHDTJeڋqquS[m?LLF1 WfTΥ4(QK[KXhOݹ~y(0"ySHސ2FRAL)9aZƊr ̻xogBNw'X袘weQO{sE!1#0X,6S 7,Y0 U97"9nj2|*,oĞxqAW&GA|Dc qnr>6nB܀Zo1HU3W'(,֜V^Z?{nP;, |ut sz=>XCӉ{RkW_ a-^;6dϧ l#a ; mU \X]JEQ{\b`S:@[!+T+X_9CjGa\W&Qj1}hVjP׌IE?.!AtwLms&^ݪ.6 ޛ"_coauxGh&Pcb_l>uMntG]; /׿jQc&}/ UKSY ͩTU.F%-"$,VWGG^@TrmUq&+,&!#H:F><9aOLxmsT슂܏)RI#OmQ3%88_$WP<_t7PE5Qu8u*Z.RRBlA=ZtcH0#@Z&".z}aRgը/v>\1sq'2lG_CJltCVtL^Im~;貳_V[#+?}Jrе{x>#HDU݌/dʾy㘐{-pr[3({6ۈn*S!q)FSG*\p`s͡4߸zB80ڠMtpJ-y+4߸Tã :}B7z~vPyvIDZcwega)#9xRӛ սM.4AkrdY/j~N6_sTʔ%wrc~ 4K ri&J,%bR(-L R.Di ~)dX !Sd0hӕBӌjTl7F%( 3&:QphjVT,oM)9 p>6dmS##'ov')qJsgkԿ_K{gn]J_S]MrU}MVg}0P: O{%`GG_.F>ȅ?J3 mc ~|cJ#cxǘMN懖EYP{ dWk1޻FzW;r;F9ի x`Wk^5EFP6̱Pk1B^VM]jHTė2V!8eBrQp'&S#E : ~s$*XjmpF2> e)O;J3fSRn! gDVXԐi+ j)rq$%I.$3>1gx#jU7Oao.J!mTT#]e S5Y[}jB'ByU8+Ͷu|'PfStɸ;XdM3ed*w2_|פ~+FrӵeO7ܮ0^_KusVE%>T@A [$co}|H'oBXLot !^t*alm@$ 5࠭1<cO$؉ʱL$U2eY̸{B'(ѕ1,\y77l7$Iy JoK+ 5T6N$H9/Ho;B $ JBvU2Ԉ$h_ 'R '3@h5dఄ `68DE2yJ]ߡ J4JT"NO ,A0s^!Uz3FgUT(g5l@)Qbf!0,ixk31#ו-<Dن]A+܁V*ϓny=>_~(O~2KݎWG٣4z $g~Q:hvBngg;ϊׇOY٪X=b}|hCy#*OQ摌]|fl'56~Tu$͙DO0H .5 c1TU!^@fDg8f Tدl GN>p+6;Tu}cծg_;_ݞn~ښiAy.PF0(LE$ODiѼP+lbM(Ćqp>F ֔?-P=k&<4?8MhLYͽR'Q~A`h- Ib 5^V͇wp2B%;EO MA3ikd0"М;\׵n1ם$hawk+1D6$,O{s?USK<׆2ѳP&$|Ʉ:NN!dĪο+r^ͭc!<tZ1,$ŎPq/W6B vlI@QԦw^!;/PX НW#"*Py^zdy!_Tg`VLsJ"+L)+r*U6FK"4WPRnv׉;)E=9#[]L6Z3Y)n~ƥW?s\FdQ_}z?o?K?3PR Jx=ynE`b_|BDH򬼸z{&" D7-N??{bt=g\ܓm^8={B%qb!kwq^q^!/>>@zx7fj Z>F)Wv]mLWN ]@˯I뛮Owy)/8~(Yj' R&~Xo0w%dU}|[`ݩR}'ugv9!ݼ$Nt0O2fr k5iq7{XJp Q!sd*_"]_]<aH E/2 )+xFrl dTh!Dȋ~?$!MևgBFAʱH.pO),/2HEZ_rz:ַeXrSe &}kM P\ 6U?)(= EhX`żyD'j5F,cV358f8"F68@'XmnWѣac m?Tb*bmiw5mK_ý%q"k\[rwED)/~$-Q_)RNMQܙ ggֲ))VR"1.EB=$RWG7WvJ\9{I|DudZ/YλslfDGc N3yBƇ:UAԡ: p ЃnMpk~VkqvvZ#1;x5eZWu/᪶* ~ptkzn|6,tSvGdm*8lR5 .A֌%$^*̔2^ȇ>A6k,[j41VZ-_Vb.&B*u.4▱X|ˬkUs 멐<xB-4O}vgE2~ h~иCt:NBqۇM=0%SZ0FWT*л#9 4\ :S;PÈaá/`xmY;MsM?5ןwϽVk"A16ZdSο3!ed^g?~C뛑m\f2]Sۛ7I~_{>.wku27ʈmB#$.fz޺ӿA93𼂒S&+Lwq4!\\ gHilTD2+c0304V+>0)/(^8 Ҫn> t@en_R~rt7?^տô@?L`筦;/)Hm轤5B)gmy*\8ܙefO/~5鄻F`.3IJ9 P m*۩SԷ=¼]yStRywפ˸]KR³ s\ٝ 2DIpebyQL9!t:Np I@bm$œO_@jgߊI!vu`G_){ՆxR/%O6#0 8roz .Ur\CH1ԐMsRYd" 75OvnکގtBHz0K_,5he4gձ–]։2qBv6꾸h&Wp1Smޮk.zzutRϞŞ 3u#b͆wXm\!)د߲)Iد/! ~}sI)xlVM/"`\h#cc9X#Q3#a1yUEKRM{b/b7Pk&hY/(8MEJEIK5i_ qlZ[)nCSxU$jCC*"5zCw~LeA*W 0-ڶsm $ަ`XIp:k:^9ke]{cEok.JM̄\zDg lC/K /;k ޯL^> 7Z5Ol_fOגKCjuހ^0rs3[ؽd~3YKTlwEU*st?}~:d6r#||ȥ%i+o}nKokB@QmfFCz֓sxIqF2FK܈"cnN_/BP8}sIv~Mv<Mq8.+KoCR%'7, r=g?NC%GP\׉qeCoxP<uy"N IP Lpeh9aDlPoTESfO]./$1K U*km*+z;o08bv4V[C y,q H[CԱS:uc=˃UwMCF[C~7?'BڪK7MF=,xlӁ%K~A8; "AXS-PbN4& X9eBABgԖJ['!pRTVapG>kc /cL c#Hisp4×,e2Eks?sdXjA,fegU3uҺ ^2[?A? NQ 5Z?%JJd=?HcxvBj 2Xps6.FĔԔv~LMF4LEYg݁;y5Cl7(_{ZKw[ĺ;NO]R]13-\!kP8Y't "XKm\}(>XSݍlWO(> /y λz E ؑmԩZc T8yfu!Iz|\05-|@XR-$F^XG2@ HLHzu:躛R >9|<21M#ED=w;#аQNa5)#S9g16{yM0}+Bbc'Tx0V1\S#8- yIlqZ9}0ӧVdkipi,bv[+}D7Omŋnn$08I-^ TYc×7TPfLj);J\&\õcpi,A@iIK0b_7={6H ѯ̧( kZg2-mШǥ.5ЃYZPߑ3mdː}p*`XCV M[4%ƓQ俹?_\}M\G%3ϏTcP'qQ)ѓHЂEBJmӱק5k% - Ò#dH)Q)CD"9[sd i\V%qPkUVJnc*3ԔBd]eS g J$WǤc&p$bN[.nS"QdiY. t]yzo e >B6ʿl:MWӋXX0{_cDʠwdvS!vGJZEem2,4 /.~yԡ&R> /.<5glRZ G -Tͻz׉50j?k(oP 圏'^'^)yiGcjBbb2td\\d4abZ.]vqh*+(Gpy|qQ\S}>&]8d.rj@K{Æ:;ųUҌUKzLQ׍??SZbҪ1vNǺk>.K2"{bֵ9錆hjc~\{~?'s'/Gø3F9tof_]ƨytV] 1F(Tpp.V5FYMt{wMq ȉH=Phnz2X9׎aN MTsy(ut,fڝwdYGX`p9][JvDآT xV`cŶ{-; -UyfBIfԺBI\B'4,w~:ۤ댿9$Ci 5?¬X A?*?? ~"yT,\"Wf% 0t/(~K{*̯.4W[=J$'?꣏BNx 2i'BhnRnXP=ₚ"P)ePvi)RO?{,pSp@g*$0/lGX¨T 9/vQ/|ݬ/޵+" N$~ E vL0諬Xd]$Kv,JjIdr AV+X*Q( Zh2=eDob ~%t=:c`+Jq.E{m9^J co6!?7Sj>|mAY, $e錂0Ǐd! )Xd]d%w3i_:]֋vIR@3Љ[} %u4}SgH !@0G[Ҷx4-?W(s :rwh2I͞fr` yW%ӱ>HVA4̞'ٜSA!87T8笱jx/ZqM qvOf9)9ݲm8!nQKL@Is{ZA.P┽V r8՛ Hq-~%l6ĘH,ө7 +A!V/^AKpY]it r!xpn0?*|x| P.4G)^RamnInmIVhS&[ s-W:ks0_̞&se90(g{QY])ƻ2h(*nCU'$KNYi1,DjMr#Z6e k4Y|M[K^C0*t"$V}&p92lEVE|W#W q Eg1i0St-C-f)=`9<k Zִ`էaq'ܔtڃ(!\P{E,f멶d(ăA&49S9j#9 Ɛ :p$f">}0 ̇K*0ɯ^A 0T?w^>^,WU"纯o_w&rf;jYOu LRL=\Q4\R%u|rwP%y;BH@ ~}W_imyNjTve2&+j f\*^w  ^P y`RYc39eH`ʐk-Hc-rrߜ1!ԝa/#ɢ$f#Cb(;JfcG. ekz8fHrx d>4Ʒ0/ E蚔RgW299ANs; %KFEָE/}"pvڊB%foLRetU!w`!n) X-p"Y#g1Bt.ӝSUhR&US+r-Pqsu%ɺ{sgr[)#($ӎ S aS/tC4fu]S}aJ:WQiZXt=sEAgx9".|m( ^YCExDXDvA ]F;yrpI#[.ܤb |Ҙe'cƭ@9dA!P[[mJqƖ1A}OjO8@\-C >qdFp#Zf02(I 0aaFva$QG\^ Ms!B<<(M"Ld1~D$to-;BlsD@lb(Xؼ# yxTL2)Wz֤ǰ }wmFL@zs{`4`]t@.2!r|{{)垃T13JʸQL.9bHD^99: #2 sp{R('4lJQz6u($sMq1.ž 19# apXbL^pxUAAZb K9(0UؓA( 1<Pc2 d6*o#eŎVwhq.^c̷o/_zQ sjS$Ѵgz{?ejdVN/MlA#Z\(6q`R)֦cnqͦA jJԴdi}1rV*K]%a?{}:~\g:f{"ut@9l]#Ϧ/խSl03γjh|ɀ>2flZegogef<-I@0Xj~־}γUiQ&7~70|H,W>qG4Y;=g(P#XuMA:^~) 6ԝ~1!Q蕊gy/7t&_xٌzEcJSsSgӴLv-+itf!^G )(Ua f_ i6_lq.L>K=؛Kx@GVaeH? F<X}娠>3}E7#ͳ[X 踐M͝FFM!qI1gXwLJ[j^ Jw,Yo/E})^{ N2QWX+fOٻ IA6;^\GDF _Wj]u HI⺿'s^jJL_~[Ϙ>֓T^ ?z ?ܔ ^Y%rM[t}I-/݊E$UoS]بo-cշZ:-'r}geE77S^գleLdv8ߜEnU.TeJfw,ܳf]ՍB!nj(zm6wdvzkƹ-SC 1ج =QurvH{nyߌ":#D{'ؙLCߧB\-󹚚~u<:3DA4I$4S D(N}f!B(x 0o8w=8$B/zأǔBU0:6. 'w.0چ:!92`YsAP +ɼ. ;( ]lh?=w͡H.j60Mvbjrdlp\b8#Ls'3d)"{5K[or-`9I- E;]&Yi.B0ǹm aUy>.*s$ZxhM6}%+oEpc1S2)j2I0h",Xe&rMO^p8˱0)WV9ʕ%V1j )qǜil49-6O5b" my+J&/At*tР!Q`!1.aTӸ>wAAg).n8I[|8sE\ҖyoVhhaq U_}p}7nIi}7˼=x2CykS(J蜴{as[xVlE=ʻTgw%F]%dqj%֕'%LFǯXDӝ[n1C9`ZC;BDAuւj@z]Om,ӽV!ç]Vq`OoNi;߭/dܣnM|$oy65|%iȊ3G]m9(,]rJFI“WtaPtG[DHj/ a$%$(A dDiHeNSc*L )MNAt}1ysc[ JdJ1RBro}PkGֹ5Zjr;md΅sPgJ" -^c\8 cT b򙦶Q'n^bR8v82nhNQ|uKz A]!@%]@{ɴ0@ NͩPAD+Ƹ6D+*hviJ,2ZiBCq)S)l@@ KEbPZ(p>n%cސ.<m޼i&\īԉLfϿH}1#yx^\`hx/7 YXo%oZi"Z8.wKm|c~y 7;;]/K Iw ąDC1+QFdĒg /\ 6zDiJg/0 MeI`{)JIbj[So#RVTϾLRaNG˩hNyǑI5y鹊2A(hŮ]_t\բb=64 E;\fi)ɞ$c@h@;%*TW&۩S"ܔT^79)= B޹lnѧtRޡ7kˢ"_/u(zvWh+\ʕݢ?f(E.ꈊ>ޖn1 R!Qպ(L{]S;)K#n&ZM7bE\7!4Y׿l|wM: >>w؜Om&R1e:)alVS.iqƫ*\BQlqVu(7dt%47K)pl3H`[%&p)$Np%0+0Qu\DP]Ԩ#Ura# *[i!RJ jan׹m_Ȓv FK{LDwKې,m m+(: d'\N)yr ځ`0䥮^:x뢗*jG78F xпMZiűv MڑPv6~!D y=K>fPj9|K_W38&Z8Us0Ƒs1.pVHmr&s$D Ǚ$M1PN4\%Rp.C?: H4:YKxiH$Y'ްYEP "9A+@!rpY۰4?ө!ڐd:=oHoU^T`|^TXr2:QZ6jO,+|훧 ̗C>y|g}W|25 џkߟksq2qՙfF\.)~gbsF?:׷ @$ۣ?cDeH$ VӐ߹䝜s ?^0f6 R`Hx'êlQ01'%@$`ZWa'Ø#Ȏ?#"j CNuba%b5#ʎ|3° QXt ˺ wK408,0(S[J: UA)}S2xMjąêWy Ԭ!oe+96 ,=6nXq6&{Naؑ"D A=,ΟZ :t<٪ٛ ٪֬)$]!Hp{O磿Mm><ތWH| RGd[灱`\IyZIХHQ*?FV~:h !;gL-i1[n=ԹnnXk%mš*+s˻TeTbHC톜Y /4Ym Jơ^Mp"w%>3^:V'LJױj)+jxt dAoFbTnP0!>b2-9$[?֑$0ema{M)qZer8asPn썮9b`şoeMV^{RVhY$% U^HeeeGF >FUFLC3?G9;k?مVCbsC†]&Úw_!Ż̓7cܼBDs3=~2̊69n9 M%=&=6 # $133?"py/S9@Bœ݋!a9N|{ ")6H11"s0XJZCў0b@LN c9L(r":eH&Qj4pcV)A٣x oc$i;P AC;s͵b5KNr)p" 'oIsC~3.y|>T~g4q8'4W>Ax8#_a~yq93p)VMzv9j2ZCS7QBث|KݻL_ЧtWGk-t{k gja>oݗ'5egqv I^i| aFftFa_hnui-5lgЧ?OzZ=Y|vuI0DįDZ8t"+%SX@6 Kn^#[Q$},'q݂RogǦ!Y.|KT6#Bd(Zv4-g.R|r$+'~[#M mj q/,6ziז8k3DQ_;6N(c'W<$@p'Sx Sowa/cI[K?D2HJI5_'x߅٪P>ed #E>z8 ;Gђ !8c*x5cQ5+6TrANj=8w/Y>\xEW"K-'Wt0{o.޻֫QXWCj(QWUCyn (W@O4Wx]ft,w6uHt~Ti=^Pv% %\ tV9^Mv}u:7zށy+!ovSszju2khӋu߷a?~藍KEC2 h=5ژ ek1+@x"ŜmkFs-7h w {<r,QvDh) Ee=H!ِ48Ph`xpfN'#Y/{G è[V)Wmhw> 8AV٠Md,_FFDfFe/\lcFC"BIyF ϳyYଇ7e7T2A0&ARy*{yzx{Lpe()_[yzx¢CmmCk+\[YS[iݹCs;z2?vSw=ݹ^4*estw?>[;_דчF%jlj#)+X^jLN +Xc}*VOwx')8-\uDϾnNwNPֵ*Y\5|9~=*%0BkQ^U߅~EYXWެJDHz"&8 ".H%e 0 52z@rX=)[s '[e:yLx97a0R.#1hAyj,:Q .>B pE:1krL ńViԐyXsJ +.MS+YԮrҵh` ݞ.СAAA/Q-Pn:6 ;ҊgGbzp>܎TFE!UՉv:}%8&cK͉Ojf+/cLw7E}ETΖzfPJIT9BEru :clczR;Vhuk!{"L118{źa[ rS:6֭ 贆Ķ[1֭ 9D0%v<W|ju;Yn !Q.ZW"ĀoC:W^pb=Nn_vU!xYN+7*g8ev-4٪@X9Zz. ٪xlT<}olUNB7ot,تpd bѨ5{҅ 5"m ^'ZN4o=K_D) [l95& HHnA/Tz siN п{ȗܷLڟ鯔9BJWir=>|ۻ۩aa$5l԰uRo:9s^%P640Űk%wsa}oIZͧ*^tL5F:}n k56[A`,ۙBEC3VR@tm[:]E3Vhuk!{"L.'\nzihJe1n}Fԙu+f4׺5=h9S͸;J^rH65 dP@^ NyKSTsXM6UzùrB5]m(A-cOG'Z^-U>l^ƉlJ\V ,bzqAOWͪb'TdʞSٓqɚ[ wҚD<3-Q "ؠB f>u'Hkf@ir,\[JTW'+e%z,Hʆ9UCnno^r6uiڞ[PV_BZ9Bwדs3 dzcJ(|tkͮ5|xAOIE)f Ϗ^D~/ȋkx1FdhhN3ǻsp#snnͻ_\~?]^5U!Ko. s$ X{1SHgbC3hBo/b5xViuwZ0C0q\wjiUCt~:Y n+^R0ar.fW>~ǟ?AlgnoN./{=F%ڰ $2?T^(K|QN`kTM"HZYXZ)ӾL=?)˔Rk>С+abU- ̔fԢs43[0`>JL1LiSyҤjuP 'dX&CyQ{*En-jwRc֨D+:b`K3FblR-y[4cNYX:44SS) .zWK>(ILתn4XQdAo2W7_ZʣR* Xs,z Lo#&5{k"umMy@Z2y#i\z{pC2avav^ ectu ޸WҜԪG,P>F@]ڞU9ݧ`M>8 1fj=$_z3Lza׳C+;Tϲ()ݝRv䖥5܆ a<׌ c 7m4 !DkRVڬfx)JB<@Tkwoj-p›$/nU ezV¦m(UCevgcXG#Տa`Xf7+k3)4>2_;W%VSMhb 3DqJ^7%TT3HSIQZ:呥S;┣ULFH?]nOGm֖;fz9o)s8ݺiƧyCcQ+ yO1&'5 }yC N9:dMd.~|lSK@uy O#h3^dE}T&w߇)ڧ!&Zs>=MR9Mh񉪄%x_6nS!j_ű_ߗj<ߌ ՘hHY{n& EX(•XͬHfJcli[vp]퓋?^g! ;Llfj:˓"^:sTH, `KX9$AEp.ſQ\fλj1(/4Ur6t8rI cuqNЇf-oseKݤ9䲏@VDpbOe/1=^#8ρL'̅(BIc&!e#2bIkJ%&0|'6niaZaȒs; V C9{2\+EbXVs63%a߭j|xj 帽K<"5a9SiN"o_sP?'=6/x[(\_=զŞM,}vK>?QڠSpg+Wfu@®g)6#,+%?u#5C-cI|0\\Dfzl狯Đ㨴2b{8J"`F U=\_ʄ6C4S"\~[5UcrH֊PSpD GB #<,| \*< :5Cr*xPGi#(7/ epj+P*T30i/ ؂0JRk J?i4EaR10!S+MMGjLf*mlɝ)wj!E!~k&d&f7dMVhfl`hz`\c:d~{-,t!~_?IZq, * #!zSV#y,K\OtwRM:,lH2cq(S]d HcG@ZIERY'VIrBh!8SqC%Z>A <ֿX~ٳ `bIB[{}[-͊B.{* ԴcPu3]8@a}bՐu޶@IAhgz&J^ky-/o>;^||)0~?ʁ'%OSJ@wkMK5Nȗ&uѡ׮B=/>'H&ӜWn&栔:x\l0HCsyÉPlbL2\p3:;guC ׼7b>W+wSy7͌F%ҳ_6͐ JNљR!^(!|, 7 =2o&7XX IRqx] } +r:x9 r1]..>,ӋsԐB+W񧳋`ᦃ+37-~! MbO )KIwbѐ-i+Y c) \e C}&QЭJS꩓^]U&0_kΜ5@ ΆJsf˛:\M9Fbfu1曝۫3=c 2}ȄC3k}x8s;V$7e=5%`sAE 9 jQR˂ԨD4FK u0M^+fz&|H9I^')aޖl\?QATpsOyD!ڑ3Tҋz1hKbltu1Bn8 }k\x0dL#qWs 3%?طƎg3]C6U5Rew\ok-w#FX5?60~䷛$֮ rf~7 nŞƶfxB_ xލG7/vL쯃xWd<2U;yW}XMsp3ṪU# ^C_#;tZd (0@SpvxlߡWxa }zf&owd;ݤ<5]<]}`e\s~zJO @q'.;8붞 3P<Zod`ťgb 1N%OG u7>A5^WXdxY{G7{EH18poph(丰 41Dž5AFe-:75SJ4Uƅj/s!-!RJ  FF3D`Y`6q)^kєm| wTwZPTi%b?#lQ ܛ}SrwL{lr=XbR*Z0]ͦ.R]DN_%v;McfjL/YZc2U@z%)A3V ]$̋GOKIL+t*-t`[,)+Fn߷j]EWhd%=+QJK}XXJ+HBEtb(|&s"}&jKE*EjlZ`2~(~Nr} 4L`Ow.?x}%nfAHܔNE:UԮzi OU/U4}JwܨSuڜA-ȸBk{js-kl3 _UWnaĨ`TEns' tX;rsjj!7_oS`V ϸhDe9Vk(ǒ %5l5D-]n vxslYɧ+6}l$Sj ʄcHQ4o <Lj$т 9W mT&T WYu :^;CXOphE 盢k=U3 ]n(cwc󚺏JU:7$@B_/j*V>Ԇ$5!@JnIU>]SR[cҔg]. bj )sǡz1F]nC禈 f^~A8#4@fѥ6JR D, :fŖF[Sw6F֢J@{rKp))t?Fsdߙ|8o)ֲnMeOBPr6J8h 3U[kY7tJnW~U;\9UE=R=}.󁃨=4 夗uZO6lb$2.PF*%x8.IT|WnNW rJZ%yf-v͸3ro (NJqsD83x",d;Qɽm{]Zg^bCnݩ7;Tc[ը,V'6e)N !k$gbɓhI0Et6=khƒu3KN˿O]Rq6 )KGy`Rq16!Z+Bi~|,?o_S_ bdwrx]쬭^[7ԉ:~mʉ)rWwVG4&/J\A4|n U2lWɰ]%vU6lSFZn@й8%F ́c$;--N>Z>|UZ#xP3!R]h:Ns*a;r;d-O$aB#^&u`)rI$ψJAtQ6g1pit$Bai-/,&0dtT'8\3QEN"齭Fq5F8M ְVR-̋Y6G- !2,A:C`-UHH8Fe Q2)Z"ihGw6sPRkUŀ*#)' A wh>]6`eӫr>_Voؔ6O9U~CS $}vK6Ԍ {3Y.AJ^h/]ʟijdFXVNA߿:Cu6_YnYTܳ"%aח_gCF \l0WZGC*ٻ6kWP|qAJ\D.v]F0Xh){zC33A5Ni~-5L7VLV8]E)H=`ٜV Qʕ㞷̑""e $:Z{DxUO`L[ֺ %FB=̐B*`٣-eU8dO .{jTQQSJ~bi#CZU{q )NoK E U>RNęT>Pd[P4 DU5;۔+5Rxf=a<|:ֹ4|'UJ!^~=>q<-d!P tbLl߭Ӻq.0ļZ@=d[ ' fH2C_38 sM6cQLj@rl ѤJ_"I&K.)Һ_ ch31UK*Maz N' TC (d |%PVR$WB:h&-rQ'T,D vQ5/ԴMt EZ2fr`qt鍧n~3*UrRjgzjf͸o,ONl@tYuW%Lm.Y_rUn=R -A.T#6GqQܖq$nh(fuE%TJ:mJ'Jp!6WnA.Bu!+@w&[@@ N A\CΤ }! 鬧䭪 3u$ 3@NٳrG뵦&$Gѹ_sg_gg_[.>Y~oCg_/X?tuڽM]2 teJ;`CIK-8;˜ʂP<*%x3 ʀb2F  ;0SvZ)8zRSRH*xQEQ՟rrqˍ<* oY 0MA[+)[MƬGafaOct43Tso]DR$28 AHˠw6Ƃ^U/ 4#)i+]HYG.Dv-d;S_@pO _ 1v]g!({Y_&PK/V9ۮ=~L띯34t}|8]7p쉖b^M{{y|L+7Z} vpkoYiwqJ^C.Ix[UY@~0j|6 #>ío nμa BQ59j0Vj0>ou˽'_:} q's_{ rXbۆgn2yap1!ݮ{ab`tF۽vuv4nC箛5eí+NWT>Q{eѼi75Ģ=o^ìqMG[ PbV `/U0^=,i{Wƛl?gwf jමvDߖlcsֿwn3WKc_v124Ȑm>C-`[ۓRJبdrpy:^OKsD,yt,AY2؉id<[NFϒIqbM9g)7g%ԕ_E[ ugɈ:vsͤRI:8GN:atۢ3o[%?tU;=?af|rGeϑ)2V}9u,`'s!r|r[RttJK}xs$r=n~4^~/Lak[sgݱy; 4D7\ 3.rRX0zFƦ^6;9u?E|w'osWxGjQ7^c]%my[ qaN>2'=mI7-0{ە8G RWar͆CD .w KHR5Qy/>9q5*25^*:>~Ӊ7=RR.ڙW@?vY].IIv.G%/m%7:UetI?FmK}Y hK}5pH<]waVX]x z ڑ¹l[ImKA%B(& +oC/uG`a/N1UӐ_li5ou;R")R!ޞuU6lPRz'R^ D20-__jqhpQ7;Ս-t!Z>"/IARDK1C-#q|Ѻ?=^LNu ǶA5WòiFpx' a藫4>sڙPu mA_45jkbCkDT^k{鳣Q7#M^Cpt^GmKo`!VJ~چ@ Ԍ_VeJV݈iZitIP \nu(tCP4Ihƶ! ,k >Su@j sSJQp,Sig%5rʊU[Wn #%[(奣R(,T>~9驸W)>N׉F4Cu)hPzn KPI (kDyٿ VR7.We! /棻ƒпn DGl!>||Na9b Ӵwa? ᯏFo#ŷ"ūMA,!xnjGN#Űus,3k e(e5&%xLlMSSFc?AjG_ |10|@m-K?yo 3?BD~|rҁKJ7& ⴧ|\\.//* f3_ZWO _arpȐD!\0EKShld٫A}4V+ApPt ;}&5˥Z5WWp*AG^bfnH&{|0u @)/"MgZf{3Ek.1_]@Xz?f`m7.ou|i%a ԭo`jV"ƪ(UVQ1[Pdj7b^ڌhN= Lj3B$Dß%[3P,>AYu-簌a-3b( g&D1)LK9*c|tZ:S \ƥ@ԞED "vߟ [ WC4dReCJ+noHښXBLUIT/9 2.31j>c.k, ULx0j 6mkMwo~|s5+rJkX,WB PJ8.gMfNai1 tV`֚2GGJ( ǎ 9¸oNfU apVrZdbIxOZq˩cL/1hy&aJA)ʼ"@FAqa6]oɱW}I6P}ac;.BbRҒꡎ!94{8$gS]:_A(yoy魟. }Q RTS%ޜLbS;?fWO]\?^?ru; > f$&ߞ_Q5Z ,'딌|`֘z *pdj#?L1g:L1cXD8 F)sF&u8-[t1\P<:&?*E)تvLoTWkߐKlY |O˛$%>>a]^+S,"g ?Ę2UP\8!h9YR%=m+(om>D_vPuG?8g h+$r*}c<:FTt$yG| %pkV6\,p}}JXajcdVHkWl:k@o w0J, CHU29|sRrQ?\qlyqµ!=DUe~LO˟F,LcM2~y?j]憅OG˔ м:!ZX˰ i-]]ϡ6wY!7/\!rvXpZA1nKRD_G$E(= I}J)hTC#uoEWcwB0%.Ka<_őBl$Y)VVMUq o'VS˜\( (:II:bls(rdNDD*nʂN3rv#+oӻYRk;/m0GS{*F 4:Jy Vu&4Sl3J:%0*S h5 Re\,*̽ne3_٭y%l$\)gzEfS1;1('@(4h g=h i\[<6iG+){cy{:ϐmGO,^`b~]o[rmxiw^]m?ɏ˫{&zsnf ih u:bV!'k>pSDfLnܳmc!sH&sȔ 3SɌ]囃])@4wM2yF_zl464$3~R]j-X˫-wu;/*ˆhUp aVhb B T;Yvy^m> IP#Nv6G|GS F:(ӄBb߅Ú|\v.OQ|f5@([E5`n](<)M)\1S 1T:F~9wDb K4NNé©z3gJpbۣ)!HAjfA:J֌I"TQUiJ}m_ȕS |;0sw؎iJ`{3+/BN׭y0wREǽETlZ+ìC9~62iH.;[+%wO'HV|r,$ TAX3=.IAQ&~F3Cz*RS1U<Q0&le,]8ERl7LfE ; X/ tVe!|._.,)A)n RL ShF_1̂,w}{d}4Üߍ(1Y<+;.A.M>DYo39 (*{ǽmO>/ut{NT ёhȡ݅1QtJ_G:RQ Y D8AX́གV3m+qsl񙛛<QBzRwγJǍԁt㗍0s}T<8@bJ\ҁ7QG D'Xwٹ?_4Yƞg|㱆ԏKF]yw@iQeuJ2-'BXuFzV޻q$<U$58}X̖D[jVe6tR19e7]3>X= CFqf_`OnUXd-eaJ׷3<ޛY#ү?][0ݕZ<, &XX%*?\ñ'"1v-B !~kw /=ص]]4 S#C-nk zȻ!B5xǫ}+,n8䀴/(v蘝&v>y1m1J) ZX}lL 8teXN޺X]O@Q&2wt??+4)e4=4Md+L7 p,.¥TtV-\-. ՑciSL9g(oN 19eR)gS=)H;C H;ʣKo[oSFXIfT>iV7m~ s`cuNjw#FI +lv~; s{\N`yr&䚰8/dNv'8 &E`pXt4%XLHРY+NJ)I,d_N9ptrB$c"E1;/HDv[ah⠹cjے܈WvJVm24.(b>^^W:( JܣA*$ԀoPJP%+ i(K8k] .V/wqQLzUTS{]!\B;dJ(6B5Cj$S):x*Af\8x>Z#|CE&.R)z5!Z!A`B(R3 4/Cq6VDdFcSD%f@9DJs;QjM0m#kԶLQrw@u(\"b0MJT;\թRXRP5)ʷf`3JB7}ƾ7ؽ-eCW^Ϗ7'1ۋOmwRFiD"EЗ"NNc*J/CuoL6tD 1e.CL&)I8&.Q@.8I ːDG:Wv@+rOBi Oy1pdIR̔噁_ 2KȘRDKOdOrpjThe7"ͻ$%҃0`Ή2|,!t~KW+o\mIqoBwPЩ>Y،&ٗ*m*bO4\-8ov)C2@ {*`SB (6xPA+{J5atsD556GP!g 3xȥ95ǤqU gc{pu9B&㷣- @Ԏg`ͼav_uo9B7z߇}DgId5*.25^n<:6 Loep%\։VxIƱ>3m{{ c' $Ø Y99>\SQ7! d@|d FIm\%[i7k%LV#w#!p1WYgћ?~ I A! TSW$AKSAO.Y_ !N!%xT_DT7W+f<)O9ہ |gd T+H}: 7Hd})NbζǴnq!@yqs xoT'\ƣQf5:ו/J$$-9u˱5Pj΅<{s+;>}7"#!.YdqImD k1%6t'DhhD(Bf8%l by+TqRlB†Hk9~ oe뫝ZR+uQϋdfG;/4PMG9/FS\1C -]I5<׀l i eM$`yϻdj+t(%%mW֝ `'Yc _:{?E=}`QliD#1T c;Ld:%\Yk1dƘ'9Y_ѭ/͔^ZZ\EѼClĕg͋=Mf~VGs8ք vnQp&hZmait'uϫW׽.θI+5aQCŝMv$(5؂(I!]Eˏo`・o3ס.[86/-V7T:!RUuVxIr9) p bJ kcΦR&YrIʩN1JgBh'I%hA匬A EjD7]V3t"o:ZGBL̄c8 'NR2M :(F RsǭT,&.'|2QqVe"z$zh}_VW ߥ4f5 OGSH9?|n1 Y>vKn.|i a'Ho}{0ɒ_ÏÛX,9`Ӯx'Jbn1_};+)#I; L4yHo ' Sh4! 嗵{8aX $= wcα1&C7 v)IZ"Yi/fN@DNJ.m7.]q vQҥ(!B3-H|IS҅v͚P;$" k-9cJ &Yp/2ܕEJzpj@G#qsb5og ϟܹOej&N&ړ v냱|> 6=N7o4s7>\_]a"&/tA45y5?m}Nב-8e_n YOsYNȞ"?3Tƌ2k H$Bo9|`:+AuZ@}&`;)o@F{yn<UMC4/c6݌ǕPOBRa)Z#,?U^/T E5>JY2At 3UgEVml&,)R5[GZϦxAd5k!8~)PR5y h857S zԊT sF,~a7Hj͉ 2_4j|>L&{p' uJMI^fWZޕylQxq2NIc fJF<^ ȦY* 2!Oia㾙 r9!twn"X;Dy̬cqH"18`)Ôj4l| q͂H:aOӏM !w<z칣q (}m/G~.SLZ߹gi`x{1M_"e?h^O3ewb9 :8c ԥN)¦a`rW& QáH?; g0.xʝIeX3LskP,tOy[ #dgy.u_,M}E֨!/6Ҽ odKSc=2zdG0>fE{Q8>]VpƑǞZn/E7ÖOr{Un955|-MaDo}4KmYwph+Y..$4"/_nDWuh[}f֘wm,Xְhϻa}xCƃZQK[nʁ<%~vIz~7=xޘRHw9t{eƺ=\bei9o3@kDc[{3m5n%+0z: ࿆A͞w< GSVwK\fڟl9IJP+v+֨U-˶Un^ƴi|8VZ3vI|& /N:#WHx4ͨ!%`g.TPa8Q  |q}ZQų`ufȗtmkTF}^exSA Pf\ZͩKBn ~I!cL|r r4ۿ|,s*G K2AeIU_|zRuPLڒK4Q1sE(fIuf^D,C[X6oJuռݗg򴐜Zw՚䅂;Ņ!:eZtyBλDNVA9a'\+tʖ&+`̈́/99R"]w.ٍˆ闸&0> ZI9-\o(UKko|Axa,OT K-s&vRT]Nʛ*0|c(;x) |`xҍX n\;k;}/$a*w.d_DéτON= 1NXQ1|Bi<16-EVyϴvB7z38o[˷ѻFQKbAqwx6I ?/w*zީyWt*#q' VPc(K8Ff+3hgCD~wM2A22O~=rZ\g{k)jU}u!g1ļCm w`wF;[x.g_:PQQ}1˛e0-Zn\l1X=v_j[t-{U]ه !i똅O pS[u\&`ƨce 5$3B(11V*3bcWŒx^̞̪XC0("Aks.6JKE8 ;#$RpE1N‘0sbKIE U+"B`N&1I@/&")sD2l婴0G3ᨨ`!\4.Ӕ Yb)A!:$$(qZ)4q' a51554A[rƎ%& ERMy/8g,d=_h<;y 9}bwo޳6n$W 9]! ^f?ʒFLWMMI6)J=k`0-YWWUX^% q>/'g[nWs 7)ў$%^rw~;x}a}z449h:wpΓL-g"#׏ I1,c4X:^lߗץdn\s$d`#J|\ y$ L( $'&dZ_khfY6LDc!ySZ8$@*Y0q$%{+Гf ^ܙh:E)]V"RtIٟH`s*z?e, 2Uc> -^$ͥ!XyEM Lb| ѓZB% =<@1P%N' T>'M(1u)#G Zٓf_!͂K >%R99n@iK\ۨCiJˠfV"(_!ʹ9U"]{dhe,hUyga A?{WF3`5ˮFsPR@S.܊yO߳t21b2S))%E}hƱ;mCL ZbGOF+afv"UKv@'vnB)mvt9 PZNnI$Tݲ"@{h)g00~ծeІS VMqN|G܋khi7Epz:)~|:đD̜v5J䲐67J0M+x|,u;gvqIb1{wq HWэWߝ-U85~qCVt{jT Uܓ/z&әUV{I=ƚjsG H^B DE.]p% V^S'V9?k-ysʎz~{GŇ){\lMBur3[=:KjlX|S훿YkDmӶs? G83N!m:NEhi!B%XXL1QCwM$ӲUeSj&`hP G ̑'pf/HI1wPR`pDKŀbUr`Ͼ,/{$;ۯh=*B +{٬6{U/JAzST6(w* @PzY ֲ;ɖ:q2ͪ%%oĎNJiS?|{d v#Ni ^\<^ Yj?奩E'O+ڂ9X0wy SH'5[Jn6=`RHxaxaǾ=7{h0'E3ꔕ \e)(!hrRKc溄u[yu&dҋ=}zQaitYQ7j|TL]\|#s=ɾq iTف~,+!=@b8?-Fr8U$@m{LΟL,/3H5./ϬH\h ٔ$y$XS ^îtW ΡZnξLpcpa|i4Y?*9[(Ío9 wcV{%em\{5 }|>溂iߘc>LJGW= } m1!#kv<DF Qˍ'zN{z@v-co"K%=w|8ݥξiu77O8Y Th~:RPr WL~9>dg}wRwןZ^K&N\hy y 6m0ozQ 4N҉I[Wr<&oLYFfΛR2q [}Q/oHGϗtutBjQSF7JJ6iTƀ(a*-.@1x-KJ'z@.F2 zXڊO{}Z:oVTAnYq kWKDo43\s"7  #-*ZGvu']c]ݛ Q- 6D A!e"zyIyq DDX r1N 岒JcE# 7oZQ&6tꀭR}rڅ#>DUZg+/ZHPEC;>^hgWcZ=k꽏G]LyI֓"O8ФNE)_ڶ˪j'Oίꭧph5Ԛqdۨk{9&'u`J]}/@"/ʓθus-ْkYzy6/:qLŮrʻWx Ƹ_"}4- M* !cH|Dxtt-C*g'Hwm9[ov;IP#ܛį'eEpAC.9\Osa!OڇI0W66Z1> 'T|H4xH}^t/!9Q0 11QpQx#%)f mئ/q5RhgYUrMqo1wD̋kr= L4TGݤ?'kل6Y_x~~๫Σolz..Ͽb1{wq v5h/0r㳅=|*ZuL|norZi*irهȦXwлY}Eu!~dfr@ )@Ph!5Ӹro]yOvX$Μ5xa>E{n<: lo@IJ7xՍ)'W% O~\VLRY8]vtQH(Pʒ/}:T_O]uI`S缿=ZQS EmM)w)`夵):R|"3œP")t(ŒF),jR mC)X.cLDCviƬJF<B6 y@,cT! u I w/QK2rq ʰ@ hymHG]?5Ϳ!5H}R:\*URL1 T3㯉sXvJB)˿Զ{Bd9Z-rmJd2i-I66jlTl wq>W! ݛ(@$&BhMiHkZFjILwsY۽ֲl *VUrv̗9拊&I'j&}CXF ʎ1JM܅xE ban_$|+V`M36""é(Fp ج:S݁n}"Ku ZxTS)鐆Fޚ`BDnšq[lz.m kٵnm_4MߎAkkPgeNO4#tmC%`3 w.Bd] $ Fԁ'AN+"U툉P{GMs2w6uYGLT0ِkȝ+NưR&0n"nPƆ]`E`mc@ Pʜ2Pngesӝ`u v5?^Av1/ W8;8.] |9a׆h/,qWĂ"l|D=D H sGG=2Q-_UIʨTxoQ(@34 `Y^ń}AhAf %Z,)N"\ 6}vp-5u=%ݶQαLUuc>gyQͳ-1Zn+a ;brf"L73\+p9 †Ө% nOA~.dTÜ8`N}rSk' YJ*QI2TU*?m';<.~Kֻ @؂z9w3۝`Yq^IVQkTdڱ sUBI߇569n1v0\0XB{/o8Qm7|Z86SwXtE`Yyf8w4ýK1)9.HU,2\6 MWuWp d3%GNЁT\贔qZe43c2'w2d64IIb0Tfp ~?`~͢,J e̵_VwjOJqc肝 Kg H쀨թЫxȢn27p~L8 `lr% c}jf54T=tqy6pimY5 Vr2jwռtx_\P0p8톓5sPݎq+5zajU8߁nGSJ@]ԩ^庙it6HӑiQsfai%agIgEɲ !wQhzg! */n=Ar7 b mϏ8%eYγ LZy΍3}SJDl0įK_I/5,7ؤ2[.O#~Ū'Ur̢h]9Ox\ߓTę*D{sOk<ܿ¿pGp^3L«nykMO$+T7))Rɦ-udHV܏O&JiXڸ_K{K|Ozi0L&]w.}sfOM .h%%fF9Jd>.xR>L7-줂Lv|e6Y &`@YqKϿ#27V7hm\S`ҙqr}qb̔ߚ/2Q*U$*RQqʝkHQ~4+X㚱H%sJfv@cEJO}^Ki|5n;SqT|ӖV52( 56K'>!v\p<J6m1c#1(jQ_';V1 hЌxQMJ\8™adO܅c@*lٺr·l"ñ'Jf92j I˕ b U!|4JQ+ {oEܴ%Qȩ*.klwh o'!eM"V9,.YR*y2@uJ1~Ҽ juZb4_mS4嚢u͚8Ddʹ +}S.ޒT30fs1e{ *wc&~.xYW3ƹvM/VS*Ur B}1 1wBFJ%۽lLXc ]njIz 0Ҧ<[-$/>u mxx9S9+wN_|7Ω*pe5l| .ߴxރydyuKbٖGeV*d1/nh, 3dAO[P:7|[,Ɓ/<$ cU7ˍ9)]Pqa`~!0u1+CBI$3Il _%֭4&S֦&o۾Kv[2JhU澏yzS=&gN}k)'7'ޢFw=7JgX __;,e6 ,UvO3>3t̝'иƘ'3͙ 5emmkGXzX䑷FxԳ7glb5c}fB1to3Ğ6L@hf3zE(qet33(A>9!*#Uª2J v'/ BobL}s8eW1@+eft>f_/ЊuѮ\xYعZ56.UU k;p:FЗ)HjXLۏ@/WZa~;G%z;{rcS~Z뮴҆6I)'p<9 z|s"|4C>'C@{?9ϟ^ȧ2uIצ)?eCnM.h:ד淿rSӣ{Y/_?s|]nZN%ur5 !kb-8 5X AD&7b+ =n:wSǵKL}I܎dG dVϏR:>qY-?,$|@Q Gz5 psm%LHrepv8FL$W,Oz6Udѓ{)8Pq~=h:c,53D߳zZnRP(x:V3];FK%$wz)84GYE"xiMϥņ& f64GVh> $D AIxJ*TR-u]]+*Гڊ裍jOQ+kPچ`A@9}]: ,!]+,gVfli36~VX܏d@]Ծ"TLUVHn1Z-Zq3wVԅ6u)mxa 9@)aR FZ\XҪWլaʁ7YX Gy|i5׭#=nW|R#r2*!<{Z,0` *'T,87KAa[Q l;%VRk)~3.fi3b1vڌ;m{ˠsA͌AJ3"n7fԫw\^:)DsiDw8q48q>0f#OےµD켚FZ/{B|ZT>ݟ/fyAQEсD +Q#5AQ:p -`{|}0?a~8!Ӻjjꦏ!u5]TWT*9 5 T_yOkOK?Lz=_R??g5 ~<9??g5P ?ίє['?gZ{8_!]h~wpsMub]ˋbE߷zHJ#Rf8H6WWUWS]]װy_LJ/x:By~~pσ}Gk=; 񓃗8ɧϋޠ /g^7.^y=;FƃwV W^5fߔ׃Ogh/$>/ sҕ n/|>?rԌQWRKSϯ{{G-Z~˯e7ϒ>*Qau^>𗗇Oow޳'O|caO+to#Par]«OG׋Aʌ鴔/ی^!˱p TlJ.Rj{)$Ǖ+~I!®~+NoF#ag>|u2,v'̺ 6գǞ?bo?E0K{^y}{e?/4<<#WCoOFo#[A<ȗ<'m`c}zc>>;{<@_W9ػ__O8:^4CӺE_#eӘ=Ե[_٨$4}ܡ2W,|!$qpwZD:bW渿sN4ZPi6#%$"!Ҟ 񜽝q`"wKu\ny;fK_δ ;ӟjmO;?R@$(3=XMIxn,P[8(h@F%řu`) 6jV|k:?t-T\hB 9b9w 31 &N+R0Ib[4ZNJruk(~:Qru\ \s[J&".2)k+UG/zԚHIkV~@B%vKP4iRuk ^6Pt6+ 43%(W|cլ>?QPk\3O}twwثȋثȋbݽYóȣ_s]x%zm:sS"'m9iIE5ipۦ9|);)ZƬ(]XIpT:L4j$hAUh|F!і9WV5ewEq7ūDԐr+ J&keΤ` 'h^+2AG]HE 652;!Zx[WWS(9׬Mp)P !j ] `BDʵK"ppC#)eTt=ߵ2gs)OM$|0+aX۽?b^9ϥR*i~~T/͚Go*BY]g͞3Y2Y%E^2Y%Eud-Xra Dub7̒Jz]G H*e%vw=^De\R1I(SΣ+EVV9?[xKyTТVu3.8Py ",D&e-BD]Ҥ~U @ʼn1 B,~\&;mH產b  T"kHĂc1 )LBF~[C$LV{-H%IOkQmfىe\(DhX}R@dBiR."MuLE*d" *z<ܻ)[. D Ѷ(jmuCV(ZH#g^)DH Ñe'`@y쐹Z)QYT-Y|? Mce$i:"g]u:c #x~cF:"Nf+5G2wAY" &ڤF<ɹ3Z_&WJGlD"ּr?#G-^"DMX[.eXRQ:iWR[ #"SKdy8 8ڥ[W٘ޚ9aEfEfEfEa".G٦"RFr2'r5:1.YrBbe D]I&V2`ҴQɪLX]cϱT&QQ5*"BҎՉ``!qT1WsFȵ"F0-.H"^x$X6{lV[|);VtCzsMbzTٝ"̐zAPztSFk-1q]DEC8vh4YtDx޼=wD#^`9ȭ\X$Ә-LK\A,KAؕpaCЛWnwbxE(Hgz-4@ ݸ thA!D+H:CRθļuIr=u _}?S;|_wzqzz*/rȡ"ʋj|5Km6^But1-QV3 -[ֶɓRFy!_TbsI (K))(n<[hy zd{r׮)P%)p%ofX YҷJ6{s#۾n}[~\Л+F[wcvmwKU\$א3D.@ۿfK6WyxJYJ jܠ  ]}޼sQs tQR YpN8N]F&9G-8eI?*,a^"Ζl"ܼew_SkL\&j$X]n,ct52/Ў EPc5"&%V/cmPTVDTo"݃iawP iI>Cz]#R ͱZvM%5zNۂB-zB̹^'pۮAmhެc}Ҳ JѭAߘ T7]= GH8lNRM^>4^GH͹zө AB#кpO5b4 B"uΪI5wdR7P$emcyvY*^eVuqE..r:n1`([cbi(֖$=Gvx֍hqܰn'5ۉCAJ-hYۉV '*1Dq'j2 2oH^5a mNn9TwS4Lta눚+y%g2ȕnXsɢ:V=Մ"UX7@]ӋjA6e*Y޶@&ˆzn~Č0)7 \=NCЊWfU PP0ب aۦщ u2'#"6Dq*jW@$"\j@Ř7H b!&oΈd[d 03Aw]r5+Q6zsI(G=K@}kϒR.H9@GPJ *w`FM$)mAJAsT[C#N%Z6TuQk<)UeNߏsbgo hK{cKCǾyU/Tytoֻyx?3Z^')0G4F#es+wIE><=m#BU_t*]mB>pBUwU*B?PnN#.Gq.O[('gȐ\;g©27 \sZy4_YRCݻsG\hЉ4AF6s-WCqB5$IcI[TM[UWZ/\Zޥ?|çz+q٥W&GCBQ)`FխV%[#X;< $, t8"zmrqbksqF]KoZ:WQZ]2NcexDL>z_5}Mb2c! (RGɝ0r^L*)MMp y 9@ V9,tS6M=;zn깛z^4@D3H)[<"&Y#OH Gιu켎0(z6o)$ʄNSi%$!kgj@2VzB!֤DAm+ Zp:qm@b`qu@Z(7")w[9u`VZj 6$)^m(4+=PjVSrn-eUdvFEns1R7 8_/˘՟[$qD1[wM[J'oꏞK?ꫩ_? al m736S95XF-'!Sw`F-,1٨덅;%?2%S(cF=^~zciKu 9Z~4j4r㓷ۗ'&>Wj&FT6x;RЁ֫ku2yקC|ԵC>%-;<;_G˄kh2SvNl+A:xp8<><qp-GO ں֯& }s|,C5 h$0^9TDBa^<?{“ H '$wÀlG! -d]߱dzlM$#D)pI@A @sp+ 8g/ǝ0 $)oFۀTP4icURV+RM$9qm8*m'CJCc播IOQ^ ;ҿaP8S"a&ʊ75nG+I#y2%31Y?^{#]da26t;0 ,-.wMj+ɨTl,JKn&-A_BJSѕ ip$L*K̳UBI,ʋ.ByIrQH}Am$w k*T 9@GSU(X[8aT1 a ;Jbiu()mzrc*J$!sLB$ag)|/x(AȞlskn./VxYt'c1DĨF**"!ۓ&*, `L+yT.fF0ow%vn?i5D*[ao!M ~UFjVN[@DQ,)h/52mc+$L4W䞖gɏ8ԆhVˠQsZZ>l-_ dNyfZCJQCUMV)l"mx6D@^vhLI,pd$wehNR-)vT-_p zjq;$uSPVf7}Qb&*i$*9ac\~dqhgk[mA g;#Jt|.yJ $+$G&d1K5,Nlj(dM ⪐AK)t4NT@Cg*r3ᇃzv7r>{g#ӥ_%<=pg^lr|f(N u,gev[Dɟ̳yk^߫ee|˞i}W^޶u_ Is튑Tw닷o q)Y(A7'q'?_qv+X7yry8qx ab)ggxVIn JLٜP.v:ݭdmͮ123 Hx}O='j8&yb9U< z=C8Z8r>AUt˽ Ӳ~ǒW=<@,Ft{qrv#Y}d&ο;f(,VܑkQQ0!լ$!3aMT!$jT)BDk?P(}~ޅTԚF%89yՇ󱬜9ۮ5t2w^,UUa6JwιlDy޽vϯS,z2Mn xg!Ǐ,xV.yuUkgu_ڬBоr5 qw=Q*$k=#`T [%_]{s8*1we:𡺹:YblmdI+6S*mNru_!Mf&U3D~21c"m JA8@cYVe651 jE&:Z>5 -RJ%M__ .XpfDu>x7CJHW& K syBCC6>HC(WV`;Az8Q>qLkZI ə-d$^F,67\'blUd{t2]p^gB =3Btɹu^m&#mm "\&apǹ6V"fn{t暴O9b3KmMX%8H # td2xt}c tex'bص)މ% u%XBm’i!OpbJw%m"hA,T'Աs-ydYqA-a8 D#31z:сeR[0=z"lGZ)_J႑,Jv!.f ɜV;5OȻmUh;m*kDřb7[ ɂV (–Kͪz,"M#ϭ=GkײQtb~0`PP7w I<5gۂB8. s7(SNLn pYRpBxM_ S}V$}.܇˩ZlCJ_3U(`}9L|xE@3Gڬd7bqvSyxp~A;YnjN vE&J.+R3/Pgv LHKꝸjMI~ ~][# S@Ok3R/Gal?=7Cdpl Ăq d #d"0((nrJ SzGJ//=9;_@G$ټCk bCWA|Ig :D:0fDJ'!;N7 vw`0y7h|_Q?c?$2Zy'qLYm9r/9zNsIwюCsq5޾ Dm o~}P lθA,2㺆r0A !IB0)Id&_SA"e: +p-.78+2t fkwR%S0=y H设W2ǿy*q4VxiJg 83)J.$X ߷[K}vb>r@˺,)?2]|Ġ4 ؠrȖ|.?G =T~!"փˆ(4Pϴ'. Zhâq#̮ccY!ʇmmk]p#w0kʂ-{Y ˀqau,{LY "ֱFkC00;Vi iw蚉mGZQ:w]St,,{ Da2XLV>VgYaFO)#~;{6Aі&#SAJ<y( p oƸLjm7CAPEE}Bn6|!bba(BAc6T*ƭl㏸=&wSlR֪|_H"5,X}Gs`l*WRfᢣU+uY8Ki.'w_WėLFj_6蠟B`7ɬ> MfC"pFvBM{hs1-S$[ujZ3n&Oc1ܔ;Ӗ׮&7I")i损Vgo8QcRuR }uZrVe#03`'L4P1ᏲC%#A0|6?Wp2ӃThby4|ԈIwd/+2 _O-\}Gްqի ;4q/>NщzꙞ^jgîkvxY.:,i}i so^ 0]ʰ#;ɪf킨oBXkct\vvl M ˦nlkwVc d"1܈yz|628(pQñзu9&BjMZNo#;]כ4KuKYW3 `5W-M*7G#QIӺt[l\56/[§4^ROS6as)βM^ש_nvxw@6\&%v MJn/1J0ᖳnWZd k7T8U].\)6\ju -@ku_Ӯ57==%,AbO+(j~bU U{yf\/<œk@{VG$KGS-4,)W~ǦAvXo5--?wy k k43fk(~,v22A`Ь/ؠ,wQǑ <Q9P:-]VB4[SQZl8@4Rn<_* 4r[vTTK5Q۪սKpuy&*n,Mzzg^u S޹d2&C<41UTyx{u[H-F\j*A8+ꀗ :6_zMP*5nkPf*Kp Cݻֽ0Nd}Kok]ze@\툸'=NO=`Jb~m&U΍_OZpQS@OϽB&/2zb0螋SʖXH[EVQEE~=Yϩ\Z9ojn_VTn=GݜYbeB2V1;lu+xʗM4Q0IDz3qGf 'b} D29Oz2VTG%ۋT:֊#ޅ#oS3JշO?r]✾٬"޼G}L¥) >t)E `'Fm1$l:{^6j,fA4 EOɢy*D*YwD,RTV+:2HƟyiJ$sp̍""bU1UIDһI(='S< yBh_+5AJ*'/|W@棆_3X|89t#%]*[ 4Vu׉4ޑB*I je 6'*!Cj6Ԏnf#ԵrvvF0×Q0@u!2h]&?(m\_EKf *CFE"E W~]d8_*(:)q1CT'QԖ]6eFf$gAvd)AdF-' #VoqnB[mL@0#f˱f¶(D උߖ]k"=|)=hw(tAKCOXql6tUja˴aRT&<U:@Z0)ZcU0I)(.u:!@DX\x eSDDiTc5E'I1l0 G!Pጋlnh!,ϛN``&s > &Lpn`r 90hKYBF^xF` ȍ0Jm1߉9+Sr, wsso/qk%z%:T{FRkfzD65}fA:#U2[Kxr Wu]78r!Yc8K0cxmO'~k[f￷Μ<6^Tb幦XIg4)La)+G!,`(t.(]%dXs;l4qQI p,Zb)!\`#ҙ-S2UE"Ka@C~ 8|V|_m Qpl #S+mp`W ^clZOaXp/H9 `P o?"HxPIf| z(M>C H>ۘI}1/T1p%>aӭgreΦpu9t%j!þEؼ]" o. bv;r7; by3d-7Lxͻ_qĉ?+oSY( gO5_þ[?黛^-jyˊ uޖW phiaMZ%ۣKq?$^%ʹ fVYnot)n0D{'7'+jƉ뫅Ǻ'κ'w,$X̐Iݠ!#If(`LCT%= )sq?7h(r̬̔163+˼<7EVHD߲"m;:Qȇ/sKe|5kX(P/!W" (+><^-a#&-h>kCyy|oЙjŅTNIj)Z_3*9 Yh4rւVDC`mB'n_+B `A0.,k=RFZ{AǑj+z%x:˛?ezz(֝j ?oSmy /?n r{\l`V6Vݞ?]{=/Ewv ohA;֔suy[D||]74OΜa~TL >\1Z u׵Վ]_݈kri!'I:%q~uS<-!6al@7Yf@և&8wv-!6eFn( ֭ 9q)SvŹ|Sl}q ^(3=?-o>\'6_\["P?l_Րͫ=mͻSMz^5X9JkBW W[Z8_y:Y<~]Lt?gO>Zvwh+>}>ۤ L<2m.ow% '? CQ ->Vwԓ.ͅy8?Hgv{{.h ;"QQ2H[DWSp65I(sb}"y\)HQ0O4ot{."|8l[`8hP,K漪7_߭cJn /,Te$˲qY|1H4%HU8''tDTS"Ã5$xI ב1m%Mz+ks%b~k9nhݿt4ܿew>Z^\ohE?PyCq ř7gMCyk*VIs{ShrUfF@噓2R-Wv]ٹ[=EoRV!Wؔ\nۗspv[;lhcd8;MQ)u~ֆe;ڐg4ݰ@Dܪ9yc9GJilL3L(s}y40(il&h =PD%a/Lh<)&!?`BLqm`Bz20"P_8hTC}@m0q,L>}ϿaP+1cfo_;9ݣSrɰ;XYƭkqhNK,)RUU<ʡ+ЏZlO(uiQ23HJy,KpQiNmםjcx۠ "Eq1Z͖J`Z Ř-lB~Fv@@|.ҩl"6ݵBs=;Tk D*BQS2=eڀ3e:= lCCN\ESt;֍j>ʃI}Gu26uKhАW$fuF#Ɣp7<I|G̈́Ql<-yp7=hȉN@+2l#&MZn"rJ:њBrQΐ^vTJ&/2gp+mUdN؁to O&3s;Q8˕Rh"ҽRi^ J[q 53%ʙdTJJ+223+tE#xcr3VBEM4U4IP>cݞк DuRcݦK {j1uKTn}hȉhwY<."ܬ𘢨K R :Ù:Y3xk"Si(QATJ, 8PY!,:SDR Զ\sWk x/OW=U yB 8_ⵐD$ӶuRaY+YQXixeeYVh (L[E@aP Vl5-A{%Naj1g󥚠@l !WBx ]Hi]ȹ͍DAc\,+y)'wx<ĕNe˒nX`mY:dhbkV"Cd*2-UHƜ-36G@id #úZ G~Q U:jhTS ) (%3_í)FK.!j%SvWEq(эvx{$G;Oբ#(Gv|ꖑ>C~+h9 1ȁ D|tvY,ۋL>n@PJIC&㴾qW/KX*yU,Q1hL=)< EYl&-1>{h]+/|E\uB:OGzP5ѲV]Zt"Dz)im]{>.?t 3wJD&142ͅed4iTH1RُpI$"cڨmһSgsyHN8jhW /5+YKN-e%r-UD) ˖wJڐWmWx@؀DS;2F5|IA|I%[jbxˇƖV7dgeџQVnh8ƅV )،s hUeqjwkI]ИsV>g=UŕG=tviR*wG0Dqv!(2Қ|LZ X ;7HMUR0VhMNi*BR6um֓U t'eAEuQӅQ M 9fhK\L 2Ris~DAÆWGα% |#D$!!2 _./v+)L/]SX}V Fd6%"(;QfFBto†q>lu%RpItLM4Ҏ%3&-w {~ ȕNZ4%, X0F l ASXZ. as&?2sC 7? [(}Mڠ[3,eſQh:B/>Z?-w~F!fJ O9rGU )AݢM"FJ:,^1pm #  +D<3Fh AGQ$^ Wl|&[S+1L4jeIVSy}XGϫ^^܌gkoS8WBrϯS!`j`E+ ;s=7z#,W}tcFF'aT25f#牂v۸bŵ 5 N_G6 lB%.nj$[I|UJ$'>Sw!r|]]ERY Ie!Zh=*YL|5X~cB7$ngr zXI(CFD䇈F,A$W(@ HaQ#P!ǀ:W.d~B?xI<.#8&@qIh'm!naCܖ(@(zk PQ"xe/ݻʹ6itmsW\&%2GH$ɪGƆIi"E3LKk|s͈DXf:EjRa-$2O/+3n۝#tͤ)dsǢpSN+,+ͦOM1<@*wSp%ܡ|=o/\%kr:ROd&-wUq6;S}U{\g X.=Ub@b.h@~/F̵р1+4GJR>X8%8cʱC%3,LQvُ7BCBm=`v2,ʐI5zPTs"i6T\8^]pw&c$N`8NH, 2L0J/}H$ "A$/QJb!f9#f=N~⹗SsbΞr\@5L 2"FW!O$Nc\K_EZ YS 1ƌG фHHzBJK(|N#|S 5 樳#QӖlɮ@Qg^ cO{WabFwAHO wZWxt(h#n2{_.Aw!ˆGh+oUlu]Zh8m$j׫ ߒC!!kU r vX@nmFSEGI! Eti^|0!:8~p*Cs('yLɞZ / |~C c9g^4ar4aE)(:ّ0#Krh6K>-BW]_|D+6 5;̬ 7 rg;Z}F5]1=OlTCg .ؔ-LΆ1n &Ȫ46ȲDt7s95uuhc xh|JʟNRiiw*Y4lYF殷Je~ȘyS{'6H]lsj &MxZgf=V.$tT8GZ/^J0ļ  Tke3TBPrn ,aAB 蛢O1f"S HYr? zGp{$BS/Ts1*9 bn~[Ϯ~b#Cn3DAVGXe: !jyfd &-"hM:g*ǀ$أ1e|LL|D %%(f FC > `)}U7 Z~07GaH{MǨd3O'qQj{.VfDmeZ 1o{$IM>8"arHŽ>2]j@XKڵ܏p|,5u``I z+@E`x$#y"‹" 󡄈xVL6!J0sv`o!Pzbj%:Ѽ_?24Ϟ>:8D[Ė'U%;JI84?txFVۈsa齋xj3> \h{jgѰv[F:u'@Baa$CfiB<Hs(lJ)"L"c8 v FYD-M+rVhXU iێxvSԖh+ l*sKw5s[5׶Y%>?Ub++hP!sO4\Gڃ,` }m1KLx^܌gkoj)T2ncD^<ęO~(.f5rxkh!du]aɾ춧; =CqsJqY V(L~aׂQ&8l2|t\ MX*sThx_hf%qM;;VVPڐ{H+H DMhfF0,MsTaa9'!=QO{4 C*"ʴYuF~@tpeNțuڶYg MGggF<wM&.ݒܜ> Bt`l%D QuJFM+pp}C  ~h[W te+R5%iu$18uPʓEV4+wUDb>AH֪'wQXCTtE_NG85پ9,4 ͬ ?84)N1 [֋ azɣd<%;(uF]@ǜfjB,P^wq~0CC3dDWp{7XQ8UfHyJJ/󐤈?BV'pZx06v`LyK4Ϗ_28َQ|m[V~%VO` Ã<}&RaO6FRr8fa1m{9PLRdX@\3jDVO6 i:ՐÈ8sD[FN@ (^h"=k"mm0-pDa X հu.NR۝K2_L`Ď&BލF?ߨ4z[7ѳx:6&˳pgߟan8,TtlnƳLl>_Q9$#W72^gL8ԏn:Ulå3U^z_-@ĉ?k_Dg/H,oyϮ/W䷞zM'|>ȟuq˫Ϸwʊf}`ǎ$uD|/n4[ 5ӬTGK ] >YQ\Hn.fF<.R~|p#ԖQ\`tX:Wavm4SjJ†VJ2 ί[Kn}/fs-˷Q~9eDX?*~:5]~uKQ:K&ۋ(jԗ_/4oC1x:}~sOw7~G!dNk|^}1MVBF;o6 sVohPOoV˼ӛɴx_g[~_|RFG5WJG5gB?b |\ewű_-&|&-},lutskp=w ${񹲊> Q$&j]0ԸUk3fn۲tvџ޵md.mp t 6 1PKV,94gH-ٔ))v*Er89;X麵+,@Lsi",h<(~eowH>(Qȧ@76{Ձ.sd 3x e^MFDNnuwF'/'x(sE쎀N ȇׯ{I ٔ<6 M}/Zwڎ@.y&$3A<^z:4I싇e_rn*ؚWF! 3ZNw `4, nOAIp5/^=M? Reh8ʫ)phץ(7 ky/wZBmLk/ZgGnk/fdnt=_F`}zf$HlMdy1qFz_cZ.erJa]¾igG}zkqE+yUNPyd##T|o4[,k"exYCxYD>cP=oËg~c6*v2}9@位Q31D p$3bK?a+-ޢz ;=vd[UM#jRE5io_>T7aw)xn׃kbPwFE Zx6ub{4=$]%\E5\%N؝ԩhIy a`B뫴n >jgW.ʸ/7o_^xrWhx DSNLD?pcpOJ}@(ᔏ ˋwף|u%C3n9^ϔYAQRR\} i *G B{B" * zYxUr) 8_YWn.*QM K^;yW tBogeLL %FЦL7 =|ܫBto6 ]_hc7F?aNדzU7Y!?@Aݗs:]/ 0lg=>MJތ2 +Td^lyǛA#jGBydiΕr`{yXDn: 9j;="X P1)In!16K"2C$Z\Q)c Qz9JSqG 1o[QɃ "Lf8e)uNKI0+%A :Q;2i  56mS\ށYN]}c̥{NJ#[4TW2kKwZ=oet<í7V ੈJ& j|c5d1Nќ=zP̀FʛF53K$RRBdX p3j t 8Q-Lƍ! jOQ# C[7:ն[B |F%bcLU voy(Dfo%X'ڏ?co_Uk6d:PMV,WY;P576XK n%ln4dנ]GW` 9`4P}6~ڟێiX9&o)_zr~;mcËq |5 ImkL4Q#,W`\n<Թoz\]uZ I#}8!m4y^҂%$ICNh~Ug:>_gR+o!mj$)*;(DNwW%Z>[_]I6(")8%N$: ֜·q9Uv ;-J󠠫Ui1tV+jC5pxWu?kWp98 1 @5gw75XKxcv ,6b.9 kKcQ52Zc$hUV3J 80MM\IO)+P1͇$.c0퉄(1!2F0M2 #8rH{IvT6 \~er6CbVLX+'L=,`'BF+plĘZȩDny aGިBm&/O??[5_4EަZp0o%kXKLR9Ge$s4m1Hչ7荊P4QhQ,S oFi$ьH!*|q2EOnz✡(ɿUV5V7Ӝ{^jHpLZ~=.E;:a{W[6Ug٧۫ZQ.Ӧzuy[7sMrxٰ 91U%m{۝S!TM[''cL]r\M]hfFK.u4cv g{^sWGL|)$zw&ћ%_DbD7k23yuKC?$祌3RMeoonўКAZ7OlO66cx@v}`بËËߤUʝ*0LUꃋq}5v5DR|s5939Gs#twbڅHoFZ ^!" ZrFzUH }@;q H B% Q(if)WM [U'_+br^9~Qj6N*ƺCpFT:^0J5&xJ>$PZ.Iu~cٯ~:?<@ !KS\25Λ:Y]5/-L_uD*-QŒ?7& \ĄTiHw_*j]vIѡf +*2bWHa4-R%H ,Q[2ͣ@JFP3 ?!@GI"1OeZܸ"=Ydȇ='wX]$ l6۞yeF׻~Ŗ4#$!iFm]bȪ 'dm;BYڡW^Vxfԝ#O.ю,ZQAmʒ1Hۯ) SW ⒖i.tAL ɶ+=+RP2.%&KCrs`$ͅ鈜d'+T ?RUdHIR ʨh\/`ˎmhB.*N^͙ÆP)|"ƺesDjeľa:9̱pSN,9~)/(^.R+T3!55Rp+8lH@Ήb6&ՆL(&5T ؐ4I! |d#] e[{qQ;b=9B }lBҞ-To%&NI Om3M*a02pۍ\vKr.)y299O19-O 2-{`:EfW "3ZcrR}6Yi,tSe}]4;@`ӹT k=ݟLkӴPX|5צ ?hDMS<֨nf!ɋj-;(Dtv zڽMy}=a6qWoxczYE=|EjgTo}Dc^y$ټfxnGc Z!yQCHa?釪lGw?)n(3.޽T]>C*R5+d&=Õd6m['ÈV;nЃmŋP2=،ބ(*n, ODa6GjS.&Fkq)ߌ:l /%&XXoC> QUiH[(`OۤXE*6]^V K?O|ؙܗggifOΞkHO_;5O"I#݆{}JM߸ϖ̨ˌYSy$ ,E5)RII3hm~9cw+?^8G+cz/lBz朑؆\Zߒۉ { cBlZ3>@&͎sBbonhk9JǟV GvǎRpΎ h6@4?VтLVkF]M-ShG5;̀S(W?rԝv92gbO?]晵2z{Wنg,z5~ .WwY%aY%8I%ة`qm`8mn)k>uKn4\$ƧEysi܃0[=BmޅH6Zݧ˅N7rO$yӣof2zFقn*&ݲOLYάBV.ś2 fS3_YcW$dqI1;j$^l;<-Vg<=N̆g^a 4֖ lg0 k+z)EޟRͷxcnlN~ցИ~!}X93Z ^TEʗ37T]Ȧ uN))<Z >s3g8\)mC$[5?]Jv\T>^Eכoc$7ՆHc!jQxAŽu4fŦd~qJld尒82+J.J+e@P: ֣Rz\< s_\%޳O~xx )H~eI\ݍĊ%$b$TOf3'OЯWd@ 9]Z0@{T#^uLI@9t936Ga̜Du"J `Fe-eE^Z:'sX斱+4")=C{Z遫BEĩ`B$d><]>R>8Gvl P*߇E3o?w=mX-0G*JE^#,me9\FsY9u7TP+(JP@B/M?JZ*2S5Bu] k곥i]l_vL6R:ZvjePH;Lܵ+/KIPxҗڗ FTW9.6- ;vA }֯mUa) ."kʅ+l)̢AJm)q-l7J §>»:hc6`& W)! gQ55a"qyd/)Ӌg':LMAj߱NyJܒ ::HO3Rz EJJF((rPX ,"^rSV)K> se$=I)/3K&S4^i؁^yhȶmwn|u pNHR25 =+A GQ3c!t)TliJՆdѵ]|(iZ<q1w{p:HO Sg>':'*pJPr$?75̈́ڊexҹP,%x. `A@e{̥)C@r6B+Q0PY'6狄ZTm؛[>Wd* ZdՑ, )K#Ԗ8ZsփiF=a*LfqiggxE-AaB]M76Rk)Yφ\v]ʦ` v_l= Gd#YL[ur[Ⱥ(nۈJfѣAR-m\:"6mhn/Q80r6;"[XC;I6 H[Ե9ߥդI24H&@DenڢvJ9rz32gHEKiGv58O8e]?Ɍe6 -m߇NcnkE$\oH4*|.O2rqPk=A31}t:~3\6YNA E3( ϴcXjNF4Co׵+)9?yG#v#-omJ &㱩uM~c;!VhQeCWe nCͷ;,#l"#(| os~íDW3h7[qJq$ΆŃzx;.ݛĚxt9cpDDl=5BSۛ5yzٻI`8vrrLh~NWKaBj}WoYTҲ"r?\ew/.:ڛ۟W\k{SC z *% K"Ft~vBְsm=pi_lߪ{ƺ^*| [LwJcMS.l!d=8SH6dEp'17.?'f=%7B]O pjf]+2|*+xmy'Ǭs^L$zѕ__ݩ)ݺ5r mu|ܦC#wu9}v|{q{7L}32v<m$ɿ"# psHrn&^?Gd߯eZHJTD HbWWur˩aJsڗo?AED *-g§aBV;7fnӸ)HπQȴģ^T$t*WxPF m1"4)5n6㧫[ )5)RvW>)ܡsP55P񺘴PÍjN;bH! Ŧ/O L`!74]]Ђ7n2O !NB {ϧQpXs{6Zd+Fv~YFOJY)B=VIJ ȡT:aSR~F1-)TH7yBi:C;Zlobniug})Xh1/` 1hp'"0k3hqfsDZHR(,zԄExt#ZF1邎:o>s밤ּ̊#:(lmc^m$%.J0U jߏۢTj#9@ϙ&c4'rvw*|0.seB: C:sjLmRL0_+P"h P]5,(2s&%3-+EI$%EOHŕ5^dDl@B&pF݅"wB)?GeTQɀ zS8ajw7s]n&q9}]cQ\y22@R='%Q+&F*80M2 k|Ia?R-bRvO"*7]<%s . RFN5G_|aѧKT/]woW|v,-ؿkiT`,3ЙሸPB4wt9ؿSh'{ wMHc$$>)Rg@w1R꣼wxv}PyP~Trk#]p3^kqxBx9J3^9'LK3^0p 9Yj*0c)?k (`?ş\<8 {|7חɛol Y}'Qy b *t$QDeLƩTXάt' >ex8k!O4+ CP\on3>1[biS\sB 11YC\~Z,nYԻl_w< z? ރG&*@Q_Ng :||FZ; .~!Sε=eѻ$;w T}zk` tA;;'(Mgs.楒Phhm ҨgFI5|[r4~ql%WoKζ(rdN5%g`aQPaWb?߿&wwi?NǏ.#:`n$$H'0:삃u9 yakn.Nǻ3Ts `ɍA3x'珿 { ܌Jm "5kB| 'Pa7fV|t1]gj0ԦBww,ƹ")oG ]/yy|ڋu8Hifɨl1M~* {sX|okW2& ~A]8? ӵ2͌O?pXbVZk+tw LJū'Co(ZUgHsw6mDDԦaMٍRi":?ܐT@dtM7TEkl^q gfL<-Hm.$xT$}*573DP &ygKA2ǍȸW,iRMx| Q 977as]l6ET2I6E43ֈmͲ:cʕ"䆦xSFAEܳ>~ ĹRQED|`4a\' xՉ}_Wk#>Uk=I 'AO"Xe#2 .ieF%$M$$£&M *\* KR'PZMjTt!"22"v&D(s+%QLYOpw8mKhb %V6I -tÄ:(`(|<m6Bdž6њR\=q3\,+I: *^ZmtNc Xe0ٶ]ӧ?DO$Cy2dc+DzD"K4|ZXEz==9q8aKHbڕG IE^FYDt4n. _0p{FBMC[yۏr}цIqw`W(@mcZoĥVʘȃ)Pr{#GʼP:hohZS]scwjIIg??LW C ]cA jF[yeSck)Ī.sgdEJRk7H*ֲNfn䷫i )6k'3mc+8&\tŁ(/u# )g]_iF BwyB`pR/_\ؖ"]Dqjy{fuQ" + ==\z wֶxЧ`L} =B#0p#HfJ&A 8Bz7]|9+8a| D'MȾ:I zn*}(5ݸJj^uM0ٴ7"V}="U 4Dzƭ&$5B$#LGRM$YcS9DŽ!)%gi_BGAdS "1F֧AB!p#I&.9$L\Z+mҿ^Lʸƻk9%zY_[cRPû;֏ 0Niؿc:?qg_vߖD 4!F9(4Po|*qK ejE#EڜXUQ=ƴA0((Cg* X]L+x*-^&f9g#}5RG7h M(K,GZ2f2OE 1ɸa T*jl!% #ѻ[%LcS4#P媚NZxfS00QZRK4y)%R3` `ajƃ ڇo5`6hC(d0g$&0 .IѥP P@&Sc^9MX"7ڠCL(gYiD8cZ;@ÁK *)6@tF|Äm{o#T0R$pD$8G!U HK_r'L_wn;7_\j;ufk/H~ψkn4~CLhGJ]\ ^1޾ybZT&-$kw+5u])= y,V3SQjӜhE_N&Ϛ[8r#xO/@x́gSD#]Cz?rNAT`t>}X &>{wG$Uw'j6F< Zc]aǍ#I_b}4C 2F#OuIIţYYUŢY_DFFdFE |`h@I%RRRpn6Vzt'ƨf/e暁%Toszn;µҔQOP!S`X^cύ5P0X˔] DtN^$u9,᜞ n{k-"!ӊ\a882xgJ0a@%Ã́#]yq?g%mcVVոolKS:1\)B7 Q&*5xg<+;a"p<1|I-UFh|Q-Id9smVGFmNaeuI%6 W") 83R:sJ 4ҊP[jV$%lgDFl(QEwk⦹ŀNd@ǔSy)q9*iw&ׂ!]y/S֟6=> ɇ 2g3E8ZFyA*q/;Rw]_epW WV֕\Vo1zYV/laYV/ipjYmK(H&sAFM7ʪ=)U_QiHT_"QuR6ֻz~[eokOh,~Ph?.sO $)p%|BZ{bZ=rBa<`K8M[),!KwwPpS8x>/h[gpp=^nUFG 8B[ᵑ!y}@*Kv)AWϙ:@a*| vݸݛ 5f/1`"Tym!n4)wSpٙЇFkA@)>ߜ%zry:Pd%c07EIA1:jG= 3ű7LcDS2dˤop (?؎wjXs:dsdh`àx m1?k6%:^)' ߹OĔvq~Ihxƀ!תF*[XZ!ip߭jppA-btOFKpu$}Vo,蚐(Z[@gPIAҷJlUx)'ݯq]:(z SK@נ!Q_TCn0!sMo΋RTzbiug뗩́RאdZ&s 'hvfCAc~OXA2Gᙐc=EOIhA-Po}/}R#Hx}8c9HIs!Wt`> clg)jY5Iћ9Aʹ[j@ƱZW}u~ڇiI|4R}X[i"L4N}~^GSn֍Ӊ:D LιqV1vVtes\%8lcG iÚۮԃVnƺPlrc=x/ "~H3)r+ 2B8-C_x(7L!F^7u{L]s5^fyLCn\vkaNVv3{sVp!qcA% \ Ju òPTE'E*(ӔRfD}*^]sQ=G8t<ƘW,z@Oyw<F%ˁ#; B7_^.}YJx|vk2`?%\+~!22PZl^f>e|%3n- (,Ltvd޸No~JRxX/F}c|9W8+FNSF߂ }(fo}ouq3=]]^bRGc|V7nuq5Z7r:,ca7 [kx1*4gk^Do.VTK$j1JXUZc_$JJL+\yIa$EْI+3Q>w2zH|R_ܩdR=}7Bqoϕ?],\C鼝="MlnR1O,dGB6EQ'|L,'beWD~>βA*5z\&5>Am_R2Br2G\T)5;W NqVŕCDsVH *{CR=Xy|JrGbk!.o.ټ &S`o%[Ud`R-ZJհ[U\8#^Y(SoF$/H1BA J[i5덏\xR`qF^9"J rbA ) _g:D# FqGífkBP9@~-& If:eI})?3#r4;)%( G o!Hෆ*=U'hF[%(w)/As[ҝ΢=9emjPl4(, |m&A-$:ڛif[t'\2CЂ2?:;Y!lc4"2^)T*~vʌ{)ZrBQ+KlTB*ڋP 1eBiyN̚MÙ7)SkDB;F!F;-l¤!lna15UPrlTBj/ + 6ʨ5uOj0KɌn5{s}+2%۷7|a OS Z!kXXξpMwƋ4>d%Ulyv0^= h)O]W1h֪7OpS4@-TwrtOaO4|nej]N`%ړ外d=7ilC\Lugy<4V&2{qS>63j9܎lzӀ1_XZ[إO|Ȋ~monޥKB0AS]R㚚E *7;D>I!(L)_{7}hCTehcd.FP'ňHԵD5%$|U c8[7t6 yDz~tLCAڻз23kf\ȝ§ucv L]_/;# ps]NNIICR<0VF=%.P@q^sݜdŔ>ǒ0qcq?0!XT_Cl3A&͗%s-("\P`xEC`y^EѰhy,r60;.ktԏMowD',ǽ5rtRrg'[LGK众O 4LM$ο;Eqg?B^}&v nqoiBKS:|:O>hgXgX0}zuzKuUWwueq^>]۵^M.JW v]~ݯEdLuWq?L"Fmш9\KyʕD`rH=i S5F#an>X9v6wykt[-x.t#D%RȄ&=߱"r,BL?.?.P-B-_EgR͹ ,u> (g 0Ih5'"Ĺ*B+1îS[ؤE FLr rDf,J*-mbOZiQ{ޛ 9wmwOeV׃v=2lY hj9z?/H|Շ06V>n/ ('y[nvރ6Oe>" 2pbx7FtOL  N$?=7-H^te9ĽFeL JoMC,Wyq ϟ`=ӕc>l=,T9WfUEMess+iK9U&riYU.h 8&LY`qe<9YAupB"ck! G*0CeFԹf12LceRI W"V p3" T gBQr 祐Ԁl;G,80`,ʀqLNL1!:{4hC@z4x?krF*Gc Q[ -X( P^(G<#AznVk]AjDVR0wRrX{q=$"D*b$e#gstvE`Hx$ 6D#4~۲S (@ @c9N DHd}ͣʟGq~^Jn\qYŗ~gYSޕƍ+b}`2($wߢmݗ(nI-Wdh&42!21ŜWrIg'n2t$Oƃ˫dOEYGb֓{?8OaC/~PB![0>Q㌟Ȭ?f}6{EU2k`Ʒ/_̰iǭY-o$ ml~u)|Qu֓ tnxr;ޞh*RqFřTHCNKV8@K0^I1+]*D.[1QVPͩf6$&i"; k௴)gNs9q F@ UAV9"4֪8X&ISƎr!2FƌA#֎E*h)BN*N4jTb?a%GU(Q*2bSCg˕8'~J͓Pl#Di> 1M n v%-hR,e~7n?B$[PrwcTM3_gd 聇3xEft3~2ϲ,%~>-RR@HG,Nș/=={#r芴t]i/$ӿ_ h4so~>DnNzyxsGc_b:Qghc)8z^pI7iE2f:PJ[!łp(E$ aGIc7c wuBcշ1f:(SKT<04E 4f  "IIjFx(=4ji\GHC[m.N"TBƦ!L4M-D>)eT Kaf}1XJSÔ"bji΄?7~_|G͎]2J0C,{"aSdXFRFw&:ĴpnvՆ*fWm|ЂLTZ>'IB8EBkcqI*<!&B&{&O6@5:YQ՘y8?IT/·MI 9w)E9 YA3L%՚1n5)EՕ l{QBTTy<˱/#hS.">#[E|L(a%QrG?&[%q125RD+ڸ`6q3ininT`  aBsr &}GV *1[b¤{b;,$eJ!H>6^ äV&ʱT&4 &%F9IJdіKJe1Oѭg9:$q:A~9.!&e'K@NSbR0"p" N;-(I5Rcx&X!t," ʴ"ѝe,'K 1$԰H5vR\(NIjR:J\X$/0~F U1?A-굷ޢ^{zduehU;*I9p<6DXTT 6Ti-jۋR❜\z9x/8H0K#_{݃<]lj{fkF1v7 2w٪I.~E@s\<ř .'c7~{f\-;i^iˏ Lnڽ}}qz27әpy%asAK]R4cBl* ѻT2ʄxF k)k_Hpkl>.{)ӜK`EAgHAvYjZ?nk)BިkF"? Vo+y\=p*)cE7M^o4j`z;{池-8fh؞t TjƓڞR2q|] @eufS/qUx*, '̓)m]Ҕb$bnA澅zQ3 FZu>B.ӊ6(!vth<_zPl7^1w$5CGRx6Ά3Gb?rHWr| !T״3ddC&E=vDe<ֹtv_Cݺ=Ōs,vSOtHqnTÅ9L&$2w04$Ja@Zr+떡?@ ˇUrGpXKe.MGjO- O2G;Tjߋm#+f=+XEw1s" h@2Ԋc!B(ŢՐ((i f њpcշ`c)fUF?=xx%09abb[5k5D5#Xb_Z$ʥghmJ} =VJgg|dB ͟U=~OKTXc;iϲ*>B_aY(vx=U2u_m28gߡك'Ia"CzOӎ[[~/DG[:+yܙPg MOlw zJ--H-=CZ1k~d۽I1zw{3*)Ia &cڍX$U}[jn˙#S ܌d,he"Rc QN)gs9]ݸy";*V~":^!#Uʜvʜ Z H\9@(q`&"{{%٨(M+AWvk"qpcPrGq653ϲ/ JG.+x*OPA߫h\rv0:C -3 є\D+ޞhp( (@#} mzͨ!"S0b"mN+Mkb#P8is"o*Zt *Iˎ.M#{N#$.hHETUgF'>9+qt[%&K\IT%~$[A 5B^[wC!EiG9-\u9w WVлRߍ=D$Xm2SPZXi,5c, 䕑 &lL|ŷ-@ Kx1`"kRmB=n }Lw%%+8ʁ)ޥzp0Z' 9SJ)$@8ρz1k1BV;՚͎E]N^El@"˰8(ܫ=*"CGDJ"KThEO5X>Wfתp1@f( s7wr56f9}`nTF /z  1QlRKmB)toݮ #ji7-W%ݝʼno:kU\!.0Cɋ7|1X?PwY$nd{7'S=5⾌=[^L_ŧ~Nx#zNtNC C"xv]wMhT&wk/\X滇$20x&wkkp ?[)b3ftgdOZѰ[~\gGM{x~?/z _E¿Ys/Wޝf:w>NAv{i'?{}?_镛r&7F޶㽷'?F_wL{=ͷ35U~zsSrq14i^ƶ5qߙYr~gO&3WYgw/ +p{heʞ Szf{ ϙWs;}{T}zVy8npk4nz\5rWx<#엮׳lfbz:_C\>} }͔wJv꾌㥷ٯ3v۽1PSw?E,wN~=1Tt'uݓ ^۝ oP^~|ݰ}z).޶{/OhڻIW|^߻go/๸|??r1}U? ˁ6C'{?@AdkmH /f~?^ ۻy!RLSH߯z/ ؁)gWϿ[o*2Ds?SR"򢊔ݧ߾-ϣB5Ȼ(Rė &}z[b;h]2`!q僌~==mfCM_*R1.ƕ YXj36ݺ6/>{!0”3:M"j=`_dwU30RnsQ"M|?}f<0{ߏKҬ7%¯^ϯ6z߁~ej"CZ6b(S/QvȣxHaD X?V|$"sy~vSGZi2?2,nNa F?O?\O0~Tk_q R#&a`/wEWtJQO9ԩt!|Xni2\ޏabM^ }KHn QU@o(&뵕X9{cRTV65p}5swhOġQIңO+9sV?| Ie T2ye7`/k03$0Q `½2D(KԄ Lg$am`NsyhM_Z>|l:bƙߥ~ƥL?gW=2Bq2|!3F9Wvt[ےzx3()Jd:O<:Hϵň[2cem\!|LlF-K [+qT1̢QV<*H8 ʥ!:f^"vكa9:Z.q<^ie/1&#'},nb߿]e \S>-Vxa,ȯ'n ͗iz}.LWwz_&kEN@)sA)=BST05xm rp-h ̘#:f ΂Cu!'7|dƋ#Rɴ"B {9FE%i$`d|lƵeڕ>4%Xbڃ*T%ssê0#j=eHcq*]SIJ`9DE0k>\6Pc}:dVC9'֎.bkjIPD,r/Rh@وs638v\]rҀ e , xd1:%$&Q\ 0wz, H% # V%cY<F+@>ayfX  6@ h7:@0wpԂ' a48,!F"LrZr* ;aoSτ7TxX2n:EǜAЀP7:@ ;G}uʂ)EV܃9}Nm=YY⅋l,RKSh ~Zɘ52Ft\q0T$XqH;\N10A  &)S%G6bQҀ$(( зAĵDg=xAp t* A xaré?Vȁ&VݛnOF֑9|=%0(KAথApCR(R:^( !bVXr/t)ya#bV`,܀J`GQ((ؼHp2(.a%^Ϊv}\$ T_tz'2'`1z#p]۟^/ov "pBWڀYyҭwןыOc()UF?0Kn6_Ewy~[X,`{s?4]sK4!Y_|}֜S<^t㎛bp?t5듏݌f(l}0_tf({mH8i g#CP*'Deom S$0Xnbl$O80^Yv"Ab#UcYArTDSSzrMA-yHg>Cz)R TVx=:? ŞKT}=ˈ|0¤# sNx 9W M`LN8)+IcpLwto% rT6&t] z]to?@I@mv EuJ%eL]2pM9kRFuLmGU-#y^gӉ8Ov;߈H$Zl9J'FfN㦩?0?܉3T6_B35;쭚n!%h<$2_!'ͣeM{j ߮rJ`Z^mZA jVkeI@Q {_^yiZPY/Մ)< 5IwrA¬LB%SkUߏUmD3:E7)Z|{@,Xm`ӣaAF 9s'8u_pP7Wp9f:aj Fy`t< ]Whbk6S;/dSN:h5JW뺂c3n֫k8Ɯ5&M-9/w˖/Ÿً993(Z6ݹ,gU!ջwfE@dz=Pϟ>N({h jyvk\*^]y1rtG8JKdRq0/?9hޫ ?\I^\3_v>&>x?Ki%)nbaoom@+xa>s o/c=Fʉ$ \1C0j_4n1(:Y@ߏ^O&> hn[)d0uUzu}Y:\%%s! Z3G˵_}3czഛ憨xa_b3NYoB)Kjh\[!:sAն I7/a5 P!0 h 0kG8*x^ L .LpnJg¸--/fgX'v] ~"}7Ww‣}Y]*h@Who"뚔P$^YZ,.kyȮu aSZ88>uf0Z'>x^SP W*V=Γ[0/v6e}KƆys*8gH76mڷ޾t{sG{s.% Jm4}AH Uz=IY[B. r;דy>Y 傒 bi)nf̯^İ;͸oSY.'(:d||89>#@\^:ox[{fGض6NPPHzÓ=՝O^U3,ۻ|"KFy) %_gg4S6rɏ(k@ o 7W1,Zd56(W(` lDGK]:DDfB4Erfk D;X$AiF2Jd U7zQv *"[8N> V1PC/ -0>s+?B8׸s@C6n>ٸuyƐւJ7{2}(Lj4uݺ9pu6{3!f)lM)*H':{nYB10rO-1*嫟^{5nLђܳms\ VbOwms}\E\E p})ZUϟde;Ҫhg,E TOj=?9UMgRRl/M +-BpLe[5kwS92N8YIe|[-FK FxBp0u^oV1@3)U[k/+oVj &LA޹*?'e8w%"sbrm_,"qXsb٫j-STn'UɶE9l=Ѿ}:ׄx |bī%x1n׳0" jW Ip0ɴ>1yp K像 8&C8S2fG6f0]$Z?M$km/ӹ_Cd tbӾi! [%{%iCF[4P4H[g\4*t;p[CyC!XZо's5:Υ~ *RTHC%nrY\lLz0gY}G}i]}s)={c'Wwrfاo5Ō翕e)>޲FY.t<˧s[fr8W6|z9͹;y野^7|>Ӽ##VLR]WĎG;](q".Z{bw9hOwi)ęJH0U#ZCIt-ҽҒ'ړ' J+9/+ _E d))MLqeQqZ^"Ǚ0)YN#$DvL)G.= >Ϊy=[;YtڄvS0]iWE}و&<+&a2[ie?V NW՗</Cff15Jfi xѓQ* 哅m\N/3 GMt<2DTg:3 i6Rl}0H(PG}z ,׬O ;D"@vs =QyxTZ&ܽK}t cյ;1W~ B|j";j /D(׉=VJ#JD$WMF2JN 9su^dOdA=k4brUI,li竩Z빨|pa ujR Ed+ ^ii)55hӦ 7q(!~DWr[L 3[dc`"ZWy ]fut;c]*ewajj8ށX5B]7Fu h"4FewDl.+t*ЎTx \%J5/&[x':?^hg/,vq*X?q'妹5=Xk$$ӑ}RŖ=cbUM>Z9{Ƒ6.oL.g3]dW:= CZ(Rg<)|1J86+8$dp"!IqXm7s8wUoVS\(uS-W!$+mK0^f&FcdڃDP3W#1(Mai`ZAvZJ$"煝^]`,2{|v>`R&VƾZTwGA4@s凤j[r kE *yZݔNIJ:8o.nmYJYtJ[4[CD' rɣ~>oߖ!bJMM4Nxj$-Є%9FwsTG.o.)lYcq@W{ݬft#~$w\ -v[PA:}ITcgǎ*iYPGtx⺐Z!BT_ 3.$ SfTV\:AB5Н2+Rh@7?[Xpa3d)q!Mw>m"N/:;% 4ԞpDI^{n6Z0XWc3GYaUf JT _%@#X5Dְ0>&RkӠ~*0ڞTʴE431 a?"+ՐYWbc:3!no8W}mu6`R US`dӽrݘե/>8(dPlE_˅j޲FfiR*֩<%'{sͲV : /&kL03/AP*'mY^\]̾UPn{*Ԟ&KɁXQ~<_[t59} Q_8orӖ5'$>bҵ1^}-k ϙ&RO^8ZX} @UqtRkDzi׮ iX'}L zUFaY<"˗9ۭX鶅É܃>X:\\ԯ>Һ0 <{ت&O˽lD]N<5_ ^\ .wŬ!bF )D2v|Nv%"`YӎT tN;dTt ڛ̓(i^!S{'Zӎ ^OWюZҙVцaUhuڽQ,hUa~̒'\.SApX-?vWvdbs+,UGScϻôZ0{jn7fGξ~Mi3HpvPtPXSxұKk^2As'Dy`SIKg,&ĜZ6kl洖UQ1cmVk)hb5Aϱז i|*߈,hr`#IH?JsEe ^]x,IJfA,Aa=c?3<$T8nib^D 0fvb@ƒ&٠H)vy4YhԵCBKݻ\ӱFd5#:x8Th]ICvܰU!ڟSz"v21vzթ{m#p5N\ LK(m֜$f=v),1+{i>qd9__ ƥoxy[| x7/(3amϒs´s}•:7j7 $BQN;TߢNJ83ᅕ"ӥRpr/ꤶyhnh /o9 { ]6PB=h,dxB,/.M /P1Hx$ƵNi|]FGqtNcqeE[:. %sb %e!gf.:ӹax)gRx~xvJ%;#]Okvxdoe[ȖYyqf>|`Ts +_n-ާCqC5#g/8eLJHqnoUiE{pG,.“\u,y^vɌRQrܖ PA8-nG->ůh63[ \o^3YcTd#VGCp+9fx I6v @11rb1LɔxfKNb.&Zn_1JpJ`;#Gt$zhp԰[-,uÂ>+D`C*WXW86˯WˀrEHؓLSjlnt8}HZ+o[~ Bpmp]^ٟ rxg_̪b"Mt~qyeBz{ieVg`ց' *<@8j2]]MwF-D=Y[a\֤5eMqY.rrHMΒ(xJ Z\Z*]N—8a{ .C?x S\O[3< ~y{q Ƅw?p֫W>OwaD·sYa4R\J0qQ"@jͤ)]T\:r6*;p<vZ(Ϲ*,f<3A+Tub҇{`a PT'(׊8 TwR&,KQP T3 A܃d }AX:*l5M`x\ۢ^ͣBZ(G2^;.O" | d]AKPaEqYH:$tbs~6$w.㨃$dBNcj+@l%n  ݇ V(}:0ki2 /w_x܋Otyi6_?=z@#LY \cph+41F:R"d\'1ƕ]S OW(|BXM2{i??^<3?88!Bq̙dTNYZ zL S>0`mXG4-KJKm4GD0c6e4䶐S$%e"`$8_O$,9Y7i:Y-L{Jx-XUwH|j է~|wtM`?~#,׏xoCwgr{3L>_@[%&:c֨< R2<}7aW9.Bt XPj&ϭBZs)yD!g`Hr , e,iJJC(v@klm# K75*?(b[*[v*eTX$c Rʿ n@`@YgzNVCà0EAh!jyO4t2W`/>֡ 2C0x|;yPc䂻|cV$ab_*FjELR*SfEPcRRQU|Ti!I0 ǚ!;%a q"3 #!Sܜ +ЏNKyR^c5R΂w쫂Fr#/rdeQ%-{ N1YϬBSTdž:<6P`"6 #  R H d\, -  wv$N?*y@[y2LVk[JòB,><ZqS>ƒK8oC֤Nj{D,/HѼ=蔉=H0E-D(`l^М=:Jb_n#\!M1^cH0w1lΖьc,%Ywh2Z]Y4%Гќ| qgݓYil +E- \[ %sJc|4Fqt3 A ǷEZ9|@sj" kMl/SС%Ăd*rW ^L\W]T^t48dǿo]k/Ӥ֓ƊVLOIju-޻%b5HZL+-(Ϟ`U GL݌!vKpy;b'_G#vv%Œ6qd& ݪS: ~Q.<7-\&\XߩR*8+Ogg eCW9y)sTpER c w!MS(1w{3zǷH(-<$u{N(MV1#kiYnl|zG F~*f˜K+f&wc 1mW]ūfUf9? :zQΤ@0*cLT u]/ '㜇r$5KC&y0ᙟjgGvq%E1峣"xNyiQJsr:$uH\[stkcnmm:(͵5`71 $vRH_b*yI#}aL ͵ZEVMCߒ3g@o>ۉ6y۸>06wfأՏ==0ܻ֟v;Ķ\^ݏᅭ8Wl͐E:S̘M"'r.mR3\ǁ: O.zZؗč9anQDvW%r/!8SticyWνhu付O:mjZgMm1&`[wh gGЏGE@ n3w. l;xnQY[^uKk5JHTS??鎫@UiPv_@lPmkܮJF Hohd}{\ڪPv™Jztr0M'G2Z^"GFDU t|5D0myйwZy=;Rp\ϕglV.t'ĨNP* .fgrqğϜ RӼz {-Ht\@@Q>-PTG sWUQ#%T(?`=#Q 3+iƤO]N|={=^˥\hrc1(G]RrB}QƬck TK,ъ܇ʜdS/z!d!0Y}ɣSqFbcFy 8%6U,Js+*wX6f&-.tخcgOvͼO9 8T; .8 R٢ .80]PU mFPIO*-(-!5V !7^s0\Gn$%tȷ/rkqv]'I`݊?Z1fgϭ5oj*x(_:>øҾqLRlɊ%3=Ggv''WBC i )b4B0 g F< 7.QRjp֌H㺐J č?eeka9j–jћi30oL4lY }7 Am~Q`YSto'`зBh GP䭉MBW=-[PI\3-M^.V1CGs RaNO!*[UuImG6.hdK|y6LB ֊?Z2?0!ڳڹe Lw>rNz6I\(y/C{luVq`KgЍDec3 hH;B4] g8ӤرUǸn;-8pFOgFtkv9.nclߍڧ?Ϡu&5~{|ѹ1o/n?{q#xoЃUjr ySɦ/oP RB eMO5(xq iYHB4V_mjNx}ˋo/_>IDynf;.\<_Wg/_w5o7cMu˗g#ykt_g/wjL ab_*KWϕE@ARx*lRZHT XN9; ix%\PZ( ޿ 7b :$#<~H).6 A=E|ty5>ӑt.\uvڛB8NHlsLNw|d)O4+΅kBV|a=H_ג"+AXl2w>++v=[7\>*+UL圆[C -8ReFQBIʐnd!e6Y'+ @o^ɇ6}?l8(&4 v`/'Ve+s!W|$\}0x! uX 8iA4X 𹇥7?ZMl3\XSwQkic޾iB"A8B^(M:ʘ; wG4ͥk_ʥOy`(!}ba'4(Yնr@?@()IO2yYyD?0O^1硙98Pv 7!YJI&K|&WjA3"vG}hVn:ѾnJPz;t{9cmFaöK6պ^3l1*X ZkDZAQ+NF%Qѻר3&ڮa7zPDK =xC/ <Fw񯟻{wĀ {ܯWW7:z&HI 7=DLe)r9Y&!R*Y`$}i^<6ێ2E/`{DŽQe''#>|<'Ql#eBP`/ױzk}[Z6m)J6(WJ ;6!0Ԭ)g`7ͧnr`6:9nͺPp sM7h=32<Δ=kSaEI`D|fQ H,HPfHL B( K8R.. LPsAlc>Uj5QeA2D,)B^`RmGp#CcGowe=naŀg0\$2+e0}V7%RQdQ1EO}_թSUg*SmCd{;ZuWٹwwJRʤma|0%޾s=̽-aʯ1>l|uJ[ځ\-7ň&uzF c[i8c3ZcE#v|;Ov s>XVwºㅜ!MpJ1:_icccb aS3zв'=wIO$c#iNo$kIx1#ˑhB5|pV/>OB!?O[ݜm<{68S9b9S#+Q&՜v~./_K`N#eRgI )t}~RpzH=Ip\H53{-DyʷVRE)$;OeGμ=t~f|H'ۯ_ ?G@ vῼ/W/|"͹McqF4scu 9{jotάDsLeLpYb5\儺#WYCbML'ٱ2eԩAp*#fW@IV -sg\o.}[۲_^"&4m^~x_}uM˷~bE搵O]vJLv'9Ar@6kSltAL4N+Lq7J(cQ)2v`f],afRp)VtLB^c$UvF.k|B0"T (t9g,z10)%jTԎqk OJ"d\d IlQűDwA/m pD§IQ8ƥLRLad }=p EGJCnWI=Vt:)*Y Tx5yG3ˏUU9S[-4i"<(qQr M"7C WՎJv/dT\7Tk[Q}41%˝aJm4b'0O9KV-%@6'.^FܻSHՁˍ>蘮 3> jHIxRU|˲Փ~y{oge՛sR4s0HÊMqU_/?%&&7yM` )FDZ@ ᆂ ZA]0jm'j}]>SXQ22?_¢^e7+;\QjU0{dZO2UrWcZJ 1\𹌿Ywwbڨu=rEőK<:&DYej2HtG4QNWӏ>qu}lp/ý R{TѷrLw.,2h(,ކ%?lL64q "SI"SZP=?ylXN oo-|۷m\8tgyIA ѳTeg8D5=w s]fOsgwRT܄o `Wo+飈_C핊˾ltюt ܅J*Ԏ5;nxXNr׀hEntb݅ v:vxH kY 'ϔ,@b0=Ź( #uL9J6CCmAR;_*R.[T(j/\vg'{Fʏkf.?κ-W uP-֌@ P ;&L)Ѩ:V^Ô&e*)Ys$r*czIN'V`ˬw ,!LN Ql.J9-teʣTyγr9O-3I@L!d2Pt4ǹD!PpQ,+.'!b f ?ŧLx\+rEN8x<r!ʹhŖx Xiє1ah!ǘMt\~*濮I;q ^s ͧT1@n~{>VWq-˲[o(ꂲQn@ a0Iye?5}HZqQ8c\䕉7>v.~^} X M/-ebR:8 ~asN"SO 9~j6_"#D#TP(=PXۣwyJ\Ts)4]C +*fn)#?!*vbnk:R#K-=8SqO1 { 1mP#cĕ؝2{&QFs3Fa.06[owJUQ!$ZB7f 3NnI '<4 F48҂F1f58u$GN)ׅ%=#Z 03&7}&\Ϟf+>2{XК =!)6@JjW8#vm+T|5b/>l}]}yťg qnH隼t}}b{̃t\,9|{+$0hgy A;};JޕV)|(ko Kɍs`֘^bo#τE>AX05`N9b$Hn:7*nE @tԈc-EH r 6ѣA9g$&cpԂь*̭ǠG1BR9U5 )+'.(/ L@gy[aǍ#IB fgh@=03,kg^6 ْe}#lxYŪa$wEdDdf) =W˔l&,J}(ԟ~x?6.d0M60؋\Μl˫?-2 KQaX0[`'"rc>:a=s![6)zuSkdzFÏDӀFlMp%&hLl7bI*4V2#,SIX>cdirE 3~t`>Lc+`Ӕa%' GSt{_<ӳt2{O֐ X 5YCO1-( l I / KpD 6WzM(XQO+Rg?9!DDU^]e?L)7F/Sd0տ3#p'Ж„s`)ʸ$*dÛ.\_#J3ME[&:@ʔѡ=ǘmpImYEtԂKѣ]N&:$q%.:'=sс݁B)} bIj-)B^ayl;yL/0ϥ >(POܦsE4=]Xس |!5-??'aߥ$컔}I`39:b#s-C"p3:6,ȉ*}szw97ϛv<0~ut]~21IJq>lgsfi>g6}xXT YE: Z:yUp؟ɢL'w>lq,ⷠfEXO'\$_^l(O`;A^^=~mr/W%RE0w(w@JPF$%2:.G"8EJ68&)TqG _nV AH(_&RSfa4 C`n35;0O˟\[?:$R޽[W[eΗRA &e<"5)CdVtLa˜; CR=e1[#?~>tX吰4"ʗ W{vd-Pg MmYuZ_կ^+a\2Cоf-)abhd՝FV ݍs >U,,nk_WC*6qA˄f)v{?knB5-\0_F"I,^a[}Gt/<㿜[a\~O_|x@%h[V㎠[?\WejS J(EE$LP^D4qBqL0Ou1i c&~m*/ |L07^!@rŕ!ѧ67yf7ھݚ7̹߯ou:ȸ3[Pl6ڿ'3x'#XVEM\zDVY ӤV;&:`HUTL jA{: 1K"MւbFMӵw/] ۞=:FĮUU\Ju:LU܉Lanz8,N'1y_j,9>d/$9ag7zl8_-_c"֮}=s}Ls(O/t0dz>i_XPvnwIuy)kwdz2YA">gaZ`,U% EpS~mX$V(5%z]ʮǎεO<)c3k> ClOLއYɣރ}>Qޤcj1/ _ƯN[b?8G[o;@v8}]h,)jnnnrժ;BUg0״88|}mco儖H~ө!"RxE:56{BӉ*"-W`ұc6?~?)E6/^|4}1EzJ|_> 4&tqz(06>·L R2Ke8H%YtbjcHKc@x5y-$!t׻zM?NRe_JD(t"I0ZFJ-0 '&8 vQ#Q{iiP( ,0bm` QmA |"|\d0[U0@0mB76 lъ:>H ""f7X溕t KBkPm#,(C^n$H8C7#=Suh\Qˤ&"#}b)VcXyꌂ!C ,x)-kr*d^ogϞe%4 +iJ`x{!5#/i0$J~j`SC!;lFTHX^^VbHJ%$>ń(*+ Dia+-V 4QkMt&~z |wp=vIǞRRW @K8M0 DT@0"@*D9'L$qލq$.UNm$]DV_bj;=C.n?mIu&U]"S1/8p4^gHSl- v/x&VXSa7d.^"+XJS`ʪԂ' WKc<6.<7+s[X-ޑ!h3EՋX]C]YxE`zI̅&fnJ6zdgÄ^T5zzص}tZ)V;ŵ;HyD*- >f%vni 6U3Oe//;gd"HŴ_K_K~m>G=1x&j_2 .| O !fUN 4UVe?y~nSj1d>jۗۼCX@'_]Q5Ln>lݹ[3Xp,d ~t=O|3J9伎+{Ξ9rcsKbh ZzzәWWDSts= ^%v>h);}r(bfOuK4N 4joFraΟՊo߾AY{z_I"jAE[`HbDXUϿaոyY-$`5:VuPFVI(ah1Rl:G U&/9v!v:q,hI.Zzm7. մIq4|mDk$*K5mj)UL+: #x,tІSJ,ly!IeLqiL4.9q`isy笴) 1 rZhEDoF#U"WhriaN=JG&jݓ"\mu_xFy=\_S JU 3WT.VnY6Щi!8`,o~: F:܋64,qB &Ǥ 3:fXQMUkUqJ0%^ufJt&cn(V+mZ dkSWA{@XOcڸ8OS1wfBhEh0>}~^O%sXE =С`U%ky\GtτcQiQlt\2/:您T`%ξyY%Wy`}R8V3[2Έ%$W`Izq)5h=݀XkpGqQ?{ȍb6`&ރ A}8 ^3x<^L8%Ynɺ SůbZ}䝧k}y?[/%d-Zܣڈ+ γD]\x9Ut='f;ʡ}Mꐢ'jXFi$.Z[ZdNwmَϺ22ݑxsQ9HS[&iފ;vp?6 Au0섌^+)S(f:< C蘚-:Ͼ(˼FQ9K=e_*lݱ9.QP ]ztƆ71ӹ^qJ;b3FlƸ͸VGP2qg mm|>#/x%t0x.>ZάehSH)X (ტL+,u̖L&[FP2 _f. \f.\P\'?A?~vp4)?3dqtJv7A0V`tGA31en׼t)5wLh%*#xA 74`Јd&2&R{D ǭM`Db4 jZF*Eo8*p,N:`RW$*0Kc,2gSgzl/毾(H9OIA;wk[sFQP kz(vnU)i>$jB2ہߙS:tWj8sW;h<'rg}?e7$ bά)_x+O,pĺ:2 p; 靉 ?ޮ/rr0pg3nٕ?VsU3H]8N}0526f}`c5D |Oӹz`J^KKu"V*Ǘ+Swz&\}6 ?巒 M2H$9aka>)I+V] 4ʠvWR SL'$/S,jhD6kDǥvCa wEVQi:hސ{3KJT+|EoFeH$o'և&N- O9*/<?/\Km=v}J2@C]çr{m\(Kd@*.ar!d*qyWR篛 ^B AM$wb;FU3*+_G JtyL/2CTQRvY.(iJk*Ė!vNkxȇ<9i "ƛoƀb rHBiCW DXsMR]Ww'dZEoO <.:j\# nTֽ+pԹ@]"-4E7)ڦwU_ Eյu^4=^з  @f^Y;ńƘ4)8B4@PwWaPAH{ԛ½ˌY.Fj4[+0ᜦ <"FxX˜Hϖ:eX`4r^r`F)qt39(%!ѬpS6(EM,lbQ&?B.i(!"<|Uy{#媄imQJCd tf1FŇʉG*t1񙥎R Z+sE_s~8p8B?_c Y#V ;bz)pspP5]^E?62R=܏?3eYi͌~Kkkvl$_B@kD!8WF ypԾWr !ӛ +ԥmviDJC"X(XiQWNôgo)J<1kWequa&eR{; wsVy:r>v*ҜiAEKvSJI? Iox Šgwg87?-h%?nc7U=/S5x8O L#sT~yV(|߿[[A[lK' ]cd8)Ν @H4 mXƢ'r>Lƃ\TPRv/ݡu9S1~'K+ r'K&(OU)yMy.!i3,,zذmI[Qկ,e<,= 49yxV!SMC^**yq[FCiXh']<6ȨAWws]wz^3/N3̙8 L k4|Eݨum{:!rGTWN~C;J{ &j}ѼNrߝSfe[v7Zf[o〕7SʛFy yڝ!PBkvQwnh0QA9!D)I&4RJ 4E2`|6w4("]9N*Gl'.0+ 벶{xW5Җv ]橺3HeɶX9'rΎ5#笠6AJh (T.G fb礴u0U jH)(Jqj?^~RH?%T .ђ(K(ArLL# iuP, C AόWg{KMLb{5 Dﮃ)+?܎>ݾ\ZyQI+Ymo.^_}Q+˺\߽d :O9!jtO3ğ/@4BHB %K=,\*Flh$+6L0ZWEZ%/YY!zEhqJDNbLN(no4)#@% >:ΎCv޸wc_A|:9Gj4;inkZ~$I 6q>t'ıSSUgW@X B<&xY'#&4t 8yڗ<t)܆8qrHWbf"| y OL@{*S _K"MLH10x1v[6Daa4* ϰ^M!2&lFEAjKX p aaa1if2L!eLa0Va*Ԉ*>W) MdX:S[֏|c.)r jz( i%/j|zoJoW!# @y~UF%Ào] kōplC9w?|#@A'h0XfS4G1[@/.#Xqۮ_1#rRK>xQ\R":Cn21N0+%!HR̛B`;䅾nJl^/O&~/mOu<]>A$Bxx֛5œR |F h Fԗ W]Ajdj I&ۅ7&w m؟F<>ۿCw? I8dS72a2B!i};_XEAX1о,VbOMXrZ3̨V7@Q*o%;jA  rHi(:H,!V" 1s- :]o8|D^_pc4ē؜<Fq`$ 1Jm8*nR `y h>1ٽYvqXbH)k.q)Wn} ̴'mwގ{v@8\7dzA߃yଲrh%ZF4STFɋNHKiS*iQ.&9;Ħ.g|᭮Ser@jfv!Z;&%I'v@S#v5nu8u1d:|_\4d:_dE!H9p$`S'4)g\1H@LIRKI}z>RR،3g ]nv|?;Y{t3od^skJdω\k,4ݤ;ɸvՃɯzP.>؃Aı~%3N7CQIQյMFIC$r *!6QIrj@wi@fPJSpejud%lQ'9n'6D |e+EVD:/B.BD92U( G1ucޮ['-EoqlOf\Y߮ԡNQ;Aik)SHځ,ǏZ8'Q$%ӯnvJsQJn=G /XYad.B=beBsh.Jixu(a/>aRAR5.t+YY[ oؿxf&(M*Yu aV 9'IdzZz_l2&s[ a/B^<?Ɓ b tjytx[J݈[ T*κSj;!@ (P~; qqp' 4ԲC?̏Y{zA( `#p).m)}lF.ՏŖd8殜~"/MʽE aYa6H1;51mXO~N)ӇI$sPC)G#JAvM)wyH扖չ UNdppNqbTևS>ppQB$ρg\ `BX /~-8٩ }R2tWCy?;t c9yiDJ?OJT(7HFhĘ]Y ]QF+]zRaZQ^AY4p&OyٸG SO0xL9QߗZS,N='c۔qyqSa$mG`q-2os=;PXTq)ǐ B#/5CsxqʵIǺEqN7y R'sWmɬk˼ Z_ne f%qx:["NhA'}أ"Ij7\O^n-8-p\HV< X@mL1S`[-8gU -λڿDtbq]7ɗIv})Pye@t0]sjDg?&骙rŅEQ kTSޥQ!|y-OzjN= q$$^ciƹ{Զ#X)h\ M5uhS ˄fq0NۮAEmo 1LeVeev )2-heM.ﴭ֔'X>dF^ Ks-qИ\ &S-ۗvx)jJ{B EJB)[a@4=/7.}UWu1_ZJ PHuuh0$(֭=q*+qt8t%!BL0)IrGET[_d=}ZuD>FTvwR'ֺB4 JT%f[)QbHHWխfN9(D6V9'/I71iґry#cZWqXĕcHI$!uJnQ/flI] x全UHWOd<b*WQjӧ6w>ZbxcߏX<_33k"EEL*oBztHcˀ`ȩVa"0-D-Cӈ֮F8]Pݯ!jr*gWr3#&D=$Y' ktk  7"pP$h/{>T?QNsAƴj+́&6.~ gRЦt_s,P-nbLCyh'H|v/Jzp8ԋocz/W,]rUV}Ow=lF;Q׏ q٣n0_#O/ :%~~ aiaIIǨ<FJH*cNj+~7coOEy24 I&[n7)m؟F<B<ed'RڿECM,m#B߀ͳvH؉v^uf)&([.;uq=RvrEXOQo+4/7Ļ":99-S+3G=48m{F<ooݛ7Te77e=`3]>~b%D;x8vnz?Coz7.y?uÕԼmB̲+8Z(m`h wa馜d+vP]8&Kf+Ԋ8dDqD (R2b̃e_];ds͕R`J=ՃXf\Z;4Լ()|]# (iGY&HܘDW=8'c=]p+5s.iwnUz`ĎJ -լw̲(F̴ן%Z;xe-4YoKO~[a:O ?'޳o/ݼ@", UJ`cV'xO=@.S@qd61h[>McP8f:1XuCף]c'` N= }=7оf}ܶ^|F/\`syX,%k*\qNE?x`2,=B <ǬóKYK[-|}O;:|2ZL#Szk-8 |?5q(DqA!1>V_nY ݋[/N1@P9V\` ,D1.9ȣ-0z cgn⁐gnRT۽IRR*3껋ɪ/̈8zBڿ\'q*PҺ쾺Цk*;A)~g\>s_i#Vh͘ZKçL@*_!/:ojz [+^_~2U7??p?7{93]}?6|aeg( _Zzc? Ap㫛O ( 0T7$~ζà`ixH#3sK6oV5Sl8" ۀ)~^ 47Ӯzc&G'4խꈺV䣞ŰgŠNh}w4 Ga95ǯ灏WK!8v `@tf1NtŘt~S=>tWצ g4:+Ʈ)qŜ%s@G4q.7A/Уwòfόf@\NXJQ|%PhJӤ$ 4%*\%K1#"D,Ġg,+Rm|& U$V"ih6^?eVMa,fMl "<\kȈGR>z7H(G?-p!XiGx!RGHm &f^.9)Qؖicg# _XLUg = JcH!:}xe( ^ZB\`9i$+!q&s4]dpˋ// j!5X! >Vz&Q{C=".fQֆ6"t};_On5 Ӆ^x pT/o۫[|4#>aRO?~~2?L%貟W!o߹4h!m*Z˽?QDnNg u7U8uaW8!yt`.a%pk8KCs<p N7 f0tBi]qBA+`-79e"ȡ˺(M 6KY+c$NMΊhWcАq*D*QKJBi+mz.̐HH~3))sǷ8yꎄ!a)oËv*m{ 劒DlR0*Qw)ult:q".Fh#:CViX֊lŒW w8M6*ÈaixiFH89C!31%`#rK$v9>Qݶ-zR1v RX>5K 6*ʨ#ZGPTt̤8cT/#gGi04E<Ѽ P_EY;Scm*rm*굽D>1MLjϔQhdpګ`W2Ythjb)zORS#Z(@ fA,N9A{JC][Z`geamey_-"ŵ;Vd:Bym}g;^(hїHT! ۰B/F 0W\c0Lcú7ӨApD . =I`̏3Dtn3&} v2ZxQ(e_ÊZrdVA#S1HP?8pxTLm :785xfOML4sM U8*/JJE;۟l4X҆WLF<8=6~{i h~@,^<;J,2%xp3gq/|~z! % rAA1o~鿙&1{ |a6z9>/;{|!Nk¶% X[gҗ<>W7O(bDNbOA7>=!sO:Osa5)幷6Pj5eJ&DJGBHhn>AESKTm.k-߶xoא9XWLx^^ue|^]rUy5gBT[k rS5`榶D~!N~[_flWW%&`G_k:qUC ZՑl e_c`t\xuA6kbB9,̨/vjQop38=–Kq3Bwn{$^CXqqۥj[0͆ v} DZFz ւf;ax* Mj6T0S\q6]|דui]i`]RX6$8a^AS{^+ i#yxJq$->m1S>LO&q<{g:~2jÖ3Mm62ȸ轿y8>YJn GC Od+h۰ǂqF ~ԭHAr gֻmu%cU *zFPPɼ -U6:KM,=%7 Vd^jNig& prj;Qa\Q=Yۆ9=:lf]GãXW&cޚ͝\7l\GxM qx>ڪr &6g!5X>hw b>ٻ6r$r]d fvv?vdٰdf߯(VewR{ j_U}lbRQhɋ%Fb Ĵ#[#NtQi/wܧɯ+25TKxI<Bsgam~;[ZaԸ-{w1yuFX}VxؾQ/\|=3Hˋײ|w{ץ`k!XyvjaHY k)Th Á<h"L-#EGFK(/ Aug0ѳ%0 Ȓ<&`ٞe0N祁_<ZM_2vkwYqydԍ"q/%Rh֧pV\E KĚnSQ9<N]Fň4N3/vlΫQTyfOIn9=EO|u: vfɲ+bȧ%JQqI\XIy baD&Qh[qu !݁sOUMԹP r;5+7Wp/پveEw2cZU=Z^*X\joNe^k3qF ~[~Y=t3Cy5 *y}Pq0[ mڰŖ-Cl,5Y&{8KtOzb) zxh5s5z/tbU3t Q M@@Lϴ' VSB$7j'I,>:wVD*":O{9賗e?ǤBp΍ _IGYyHR ̘yqA Rah%C!Tչ| :oK$YP>ą&-I'@8a;/oJJ$B'1@)6 LXmLEXDD3")WRu֓&[$[+|2堒 ō "%>yND' 6&+rPZk-u8"y%!TbUaͫ f@`IdԋH!$^8hu,Cpޒe ՜2~ݯIdu@a~qie3|Hu O!;Atq @lO[Tcpuw$TE8/.#zdFEcMmQ%IQm@V!hK`ίRo/m rv,"ko|zt_8{sΕA>}}3@Fh:_F}6|Z;?5/it7-x<\c8 Aet?QݹFJs6e3s)p[5CǻYBdc/SfFnyƘoKxw4eJ=h_C9Aje,Q6[VKNjTFK*l9g$(5xLkh-ϯ.l\6qӌ^D1x2j#g^Equr~='%2º2!`-QW({_.| R u =iIHHFB3xRH<1mHg8 $<ݑ3К*[Q4/ Q3 *J%բbKUv.RRŕ/BO]gJ >WQAsGeF3uAT$*k Tg- jꀇ`l.9i+ !'(!CV;Tq%Wfs'a2.7ve]_nsMQO^w<.kt R|&%!mp ^໮!T2Fm>kP<Qm1 )b*OgN`&?cK5K=[|w/jf~g x;_s)v05'w seFK DtPWSKu`RчSg;lI5rEj~cU^n}nKg r8ږbPmϷTV\]w^._rEH8 }nj]8GvBy x}t+߂q-)czo'ok7X[)uDt&gMkj6$䍋hLQv3JAbPGtJhNyG"%ɻn}jj6$䍋LAMs˗|:,.ʕEV늦[4y嵝ثEPK?fߗU.M0e.j5(4sG ȧR9_^S ]|Qtt5ΓcAD{Ů\;ΎTi8uZv0I3i&QƓ@ew1_Lp? >)N~{#|?C%Q7.'Wԝ]IF%RX ≇S[btνKIBx7J_"ֶv`l D j%a0H%eTF3ܼ0r7CkLޢ99 ܨ# KJfW9`Zj{tb6wFEm4gKYYTDvR_V ߔ6\`i!+-Ϛ]ә~;XC^< ?E"9ӌXBD*pg.M"(2T*F m +mABu-a@avm)bܔkmXR\+YۊW!-PIB }|$"Jx(E\sJÃDF" K0Ӫ*Lk($Q/d+4Wj.oBӎʂ(J,Sj<("'4)" LN"QDh RSf2`BW +; ޝf puMQDXgYe*~*iSNy_o;2[yLtEU"lFh)@Tʖ:4O*G+L9G9GhQ)g}I^#I=J*l?,+#R.١x|\`e[[3RBH(OV8ZD*Y9VȜ4z"d |Վ=PS}2p3{cxФIBzq)r'sh$M+Mn,x{P+nZ0܎*u5MwSz93ْPV~=8CMt;-< yb*y?7XKW0~- ݠ( .I=s!)YszZ&ZPċ*x1yтdw[#i]t S8

U--n.L=\9I8;G+h QJS>~mY؍45%}PaX;t<p3!2ϱ\v{ @[(6sCň 6hlv-RIGH+a8JCg?x@WqYCp5*ΠɵNa2 u}gG/~1*zaPM 4,F("/b3 ȓ.Ad$:9oݧQ{ deT3RkRYmf)~^eXr9CCh^!g/zx_d<ב 6ਖrk#{Yy1s{])!4y_{|f-=@Q7&vrl׳Gg."NKC+-'rrmy |Hl;!\UsGx';!t 4SpiU638 ל 2-vpnu}29њZvD&ao 7!'[V8S`+dΗ*1Lߎaj /-@ލG~8O>.W 2UE&3Ўy}FvXTl{}\tu:3l_5gt%@Aw73ƪD*41Uh CmFE/2ߋ40aw p70camِߏd[nILlŧbP3P{i((ẛ>Z;(D~ xw-nwQ6 ۿjֶ"z޹5Gqg%d7A>۷ЊnBۭ}} eJM(<:s0I2Qd$A)R` h}Iv6VϹ:"RS{^;$zk 0I-JYQyE.P8٠( cbk2N 8joU/SB%MS,؀'l _Xf Rc#B\8JxO0H+Th)M.lFA&h. &=Sa% %&<R(9 ֥C$g0"ԙYѹ7Tc:#2 %X Y^u Iz|;þȓ_jeoHг9YrܥpuR{)ӫ>{),ObfVy[][d]PquTKMka|C1E-3୳ Bd'Y}gwǜh4cDRw #Nn&B0?dG!4`Sߕ@m,d'TbO5aX~#+lM[@!$Me^RB o`DTyT&"oLJBI0|Y X~`F0,჌*kbcb2=}!b\N`|Ԍ"E{p1\+Ip"Dy9S_FVg'$>~Rj.6*5aq0ZÂ1Rj0jfe:8ywfWa{5_q|°]y, X=ndza[c~ 8 g=a*dhGbm]Ӣz0-`EB o LX˝O;~ZJ ]΀!Kc-1Q2Cɽθ%&y18Ȟ9RC٬Jr`k:XʡɩwR Y=bPJ.)7hritXA) bPJ#Yc d`w\F>MgA)PzG) T=YRVʯ_EXy)C ^1]_0#ُVW19- ח&OS]J?Ža#ۇ鬈׺i7~ +}JmzIrk-K rI%ȭWr qMJF.Fs}Gm1}%E=Ƿ6F7>DƔvK0"_.@-4VG\0% _ք˚LRMv/nE jR絼?^LxwnZM4H+-2<)itJivjz ST޺]$KLd ;p(NtHG5z@V }Vts&xd3n;3BeVݡ >8tB*^^~):%?t+ٻu8:O7z?.F/g|g.11?;f$wtߵ*szS f߿&2.~yg=>cj:Ev#4RG9H#}aR*%ZPJZ=U貳GoUv6W@gcfoQUG7U V,%ThzOM?x4Crۣ5ʖ[T"ɺ?l {z"끨Ԝɑ^$'ғ|_\(tզ#+J)L&Sy]8X}bo /1dplq6E&<5VDB M/Trym&bꔖq]%?54?KRs@^V?9Q Hg VxX*~wRHeKܹRD$`#,P8B؃kFp=r.nJu Wqk1= wwŕ|%%]/~{B|oڀǸT_G],z-֌mgVhڋ$T+@LϹ;.gp&j{-x! K,bgt(ï[TtH{A̵K2DTX,CþH?TxIzq׌\ϜQ բ1] g S2Jd)bg!߂AEԑҖ`('3On(e9\0{S:2;M ElkE$&L3bK Il)윇ecz5vlg͐6 "=e2Q Q W.mN](X=DVW[~,~: #e]_[30 C4Ă_E_z}7 Ss^;tg0'5 ?7xz^C{qX /uof87QR磌ms !*DA5w!%C,9ڟWԖũ$#kV2nmKO #Υ7Ghlc2wϕN$qjgDV{-_xbNfaPR+u[ X`7qP.N*T$ϝ6Rb JDytAs܁[ 3.f/MRLNhyjf㛼}9T|H EO6SDƾ|;UMkqʐK' ~4R uښUӶ2*IUuW].&o^pLxwn-z&h=5@ԟ"WJW!G\vyA@ѧ hBM8n2zזPLQXeR4„3v<Ҙ]p:W{N`9d) Y^îĮH)>yk±֮Vbc\1D(*(RBKPmޟ4NU3*1rg_HM+Qr ߠagxm&iLcc`a Vsb\28Pr 8EgD!3Tf3lJz0" 10%VV0pӁaqDbe^beu!\JLp}8fn>obr~Mt]O>-seObS4&#!/?>-B"wϟ<>F>()?_A9LgӼd0.[Ν H|plq00sܿ aY *b =Eqv B0ƍ:Ƃ`y$ &V pE;֤f(Mq%Gmou]"5bkԵd^9=5z*8gu\H͵la4W0.(m%f!+EAbA ꖌC$p!`86Zr ֤X)v4sVUJ 5tp.z,l1:0(ՅsLBWH$ W}@<;Y+ힹ:$"+9c[_kqj"=|tulkC:aߤ#T!iWl:.oŊ૮F^=+BIOj/SN(jyxJ UX>/&i3^!>9#`_Nwalk.}74HZJ_ROyARm~8TF]P_+TG87v1q!C.(`Vg)J&^{?Ł4R4! '/]J"9-V!%jGHDoD@((9YPN3AKp8UF eH㠉!aߌEU"D&<9#IZnޱO~} |)pfވ*95>D pv(CmRr>EQpr *j4xon/ӥ)WGpPĥdLI$cpF5h⨕! J\z"^^)0eKlYpɡÓGJlvTE#h2c[D hRƏ!Zn: ?YH;|nߟHVWPY)$MK(t.APz 02qO 6U?؝-i%WmE:k;A%@YtJTQ7<:kU }Je mO*v8(X+)PC4oX,ڍ/ m}碕Z sϖ2pFyeUV9vͲ^{VrAq ז30Ђ:^+HƉZ[($FXP-Jq"kΠ@4WeN҉Iǥ%õ, PB5 6 3 8Ebu_ۮ%V??tҭ0ѡNiY^%9UC^2vXK!V\?E]jhNQ;j_rcj]=@=ضAEn"t[ˆ>yW|WN r6wU访|8*X8)B˶eGE9eHe}m7D+<Wv`WpS{*ۊ>1J"U\HG> JcȐ*zU?=:詬rFZ)̹Uy=9<\ais1%z<TnC0FiQSM1I뷦r֌QuMv%Q[ 0vӠW{Gݽ([8`*wl+z5$8=T嬩% rCMC˳A\2.cLdVњ*D@pjvP)avAjuYDҠĞ$lڇ $UHkyF@BȤ``A"!TD(Q+Avhm=N n+Q;V#|[~w0En`vO?H~\Iֳ[Pdߐ`KZۍ߭Ne|J$mmT.~ԡ t wQȑ|&*SN4h+&OLuZ~<絈뽳ԟkێVSH90 q%c=@5vS,vJ%9c3Qzkʬ s 9_ʊébANzѐNk}r%8T<+Cѓ8)ʹ5s8BC>~JQp^ mik1fnz-7ƙb"Q)AlFR) xĥ}9BW+Gs4R+L+aisiV""8Yk$.vc~ D?BQo 2n%(QikW mk+um)iuG[I'5Al͢ Ù@6V TBX;dzBF Ln 液`r2q i=ȥ!*jMf ev8";Z%=^5R„$^~s>??x.(Y gbkqV"XNRWYL;Iv%-@CΏR1lLڐP%õ,kj*ar_/ P W*d?_=V78rR6?0V3wqGR:RRvrVm;5Ԩcb,* B|=h*ȒvsqFۺe0)aðhNyoi:q"'!mJj1qm=):.'%Li)V4p6Z Dj)JR]JI(  w8x])]~8cL>8=!JzaeiKn )*bX mz3f+it5biVC,bMLR{C21N$JG)#JGgQ2_8^h8z ٚJبަǫ}qܒFզ?gr-xAǑNHk:;'|.BՄDZlۋ,E_9i@ 8tۦ/RO[Y*=\/E@71eȕA.+|Yᧈ) s% h h6`/"4L ߬K.U{Jm=1d 7]C$y?E?'慲 s\ܬ]o˛ =~'72^}'ŌFfYG*-ONf1^c~l?ەF~?:eO FzX )A[kNkƁ52e|m;X0x!?(.Xxjdoã3A^E b[Dž noNnnlk(;m'/>~?G֝n+Ilk>\} ww}ng3҃.:A1zac;& u l{!1WCãd{ƀn~Iodn\mGQ!e>1b<{%Z9lUwW$ϟNn~}QZdLƛcWaJjz6HgxO?D4./nQ=Cu"XmiDQ!.?ŻA OcNC>t愪ӷ-ɮXmb'H !ұ / !s-z4jiidk!4Fe@-4 ]^nԕZW|R䫁^ (ihcD@ RgU3bvͺ.(S <b3$뚥@$7՟!L06V p[8&pv;;U$ BjEXi*[Y?凗 oHHN5Ҩ'߽ ,3 O?6- "Fu>LgիaQX7W<*] Wνzb3Y!qVhpB/nR\au^"yI_`Ti2_PpA&L1x1ͥ c jy( :d@_;N @v5[jZ`P ljB-QXV*AZČ)Xc`v(Xe;sƂacp`シ`\"vZ7vI { 1[VoS_:P( '*$7A`rYSubRƲqզvCy8T,đ~OANa~eLSw XA,t %)fD\ݓ5 SpW/qB"%Ęr\\q!qE㝳N=[IDZNzp?$:ƜƱjRUI R.9iOW1 +:4ʨfҀ8twpêhR\"hI ތNpUĨRF,- يt~pK/ !6$7܆rrCc֡xu2ec1,:"q,[4&mꃒ.٦>|?x$+x$-7?Zꚟo^Dt+A#|N}P:gqȳo TR 3EEE)93Q%FYr5*Ha@\JJDwP!1C W3xqS!m!I0s*5\MᜨNN`j af[uփM`L$sbRv8'5ձ᜜s3>Iy+"́֜8'Qˤ!1+e@e`I{q'`^= a9A{ \X̃os#"V( 3Fg0F4KOyƎ+ bԔd|N~Ve ue ͑5.r_L/Wb(<4 &Q Z.A_m/򽁫qϫQ൮k"7 @&aգ{'f2^V)JUD$>biy8TC8'PA"Xl+X>9^>pT'$[i)m@|hvMy̗d?/l)O]B,ύaRf3D r絵l0垁ʔv2uɳ<7$vt꼥,ĩuyK 󖀑$D`RUAfyKcŤQ#0Ԃ p~;24~ȋ$UUn\~T&ٍ3?-1{¸ebD @pn "Wa3 ި˰ ,O {?_U{s} ()7;u/|s=],b`W,\|J%Y=92yƋ3G7\Q$(B>P2@@Q5壭|JPK%z /p%ѽzmJ)JַtYR _Ƙa0 0 Q]vN0P~0!݄kْr;q~4 ЎI|ŎzɎbG5 1eqzF1^HGu3 '3k?C?y 1s  9'Qh׋AUo x;Jx(z}M_e\)E)ݻSPr͌.C[/X3Eto.=1\X<ʼn Oqqr-(%խh&BJIss(sΣ#iB?5%80_>1,Ng[B"&EU獣*qeVF~!GUGx8c[c KϳK%&Qאw9Eљ _L5A3;ݥlN($ -d= qGhTuOԟ)ښwB fFzSMpO Ú>0 }l?ݕ|:Ⱥi(O]N9d@)EԎb6]4Q-m {I_*3.ж"lkZI<}_K⏽xkYoeY*l}ѮptH{9wV/6xwfҲbhXkT\": % F.(D aj+Fs.:1#z"L&r+iT.H8T0y&< t@K-jHXUEn1͊&j7cF-T bwIcO$ &qe;xĺLiYIЂ#OE>{*_J (yw Êӡa% LQ)!FJi0̍4ZqL甠,3c*J:0 [])nT H^d~OJJmp` b8w=hC wꯊ 3A$J .jz_5 &~_Us.d\^{ B'&˘2p֒p",rL\Qe(":r$vuw9.ԗRHr8(ɉ$x1Q-C'\jxg(4\8E8W^]xߜԦoSW?{CO[Ca9g[Pa?5۪S,GJ N;]%Y6bBX2&tuNZ`3W:ŤsLR&{,U+8HQtu Q, Jq`+9z KeMgt\\Psp(O"؟B#@k>}EaٚSux+&eJZ'0$aƜxoģGf.Txˬ bedQ㛋\2?s,ea n[Ah1?-u$Jބɳ|0@S%Hڻ1/E˖M.cr.f=KV '?/ry.|,Sk}&K [*1&&ݒ7+EwxJEB2ԉ>o|_b6YLtwlv,$)@qyY,Um6 Zwu{kiq`ͧeaR Osz{ף{კMf oÚ_;\`-J aq)ⲋ~$Ɵ[ QP $+Rʃ4BiHnh\|ȕI4#K3ѧT䪷A"=8Y:%:2JJ Z Q\Dd WI@~TP8rtC*EOh=6J"De +߃&R$Re([ѨF61V嬂J"UmfXRZ#[$@l:JEw5,3HJ#00jbC ն] +1OAȡh]&K.g4ޭPxfEHZ83HYX3MlmVwy]֕*A;Y/c7=j^q>qcY'ûz;( Xhv(іʪ^q[8"N+͗9Z}~?cUtQȈ1jy䍓_|*󊙵ՍH X"þY6t &?]Cyp="Դ2`e}MeJzŧq3(7)+ IkňiFF+7zC b;e4"LPOOF\Y@Ϗ__.qg- +gtͳI4ΨdG81_-9qwz7E#g`Q(2B0nwρkv ID>$7Df$ 3W\5.k1Fc]V[kЪJ*Qo==L3 `mp}-GNeF5*RTQI;8F*00ٜ6ǹLQ?{㶑l/ ~ 0 ,lnnxa|4NJ54N&5RK"PdFY}U]u )M$YL)V`@<=] ]E}e–dXP@*X<* XJhD1!D*lLg6d"<ˉ̥VRf@JS{ .կBzY,I12j4K2ƒ@1U.I^X0JP^RT}G A#}'LxVI~3u? m*Ze{&xMWE@(lH /_Bp6.F 8] t'OaTp]Sw\S_i^=~[!A76KoJN&0CZ{AE]{_zwi[]Hȿs'$m+ DjYFcg=8uh t^9& 40t__KChk狁K0g4ikסNf"umkb)5(M>nue>u=Uжu?rn!f g)Ȩ`$\N(4f2\Og 7G!vT9/k$zA향e1D9ny^R?/ʐ `mPq@ 2H+zN;TJ&Jڝu-dXsQ~{!, pda jwbؿ6ݵ򒫈Q]]n܅T`'f1ŸH 8Q$I`*8O&0[,Sd\& ,Or0x62ݛ/i]S:} Y!6cϙG~ {eREDg0")#!dDHE5x:\X^fL'sz 椣?VZIANJ8:Y_o7G%UP ^v$WߌeVKnV6T= z,R54=r} C7?,[FEr<NeHy&o0!` [[fPZ<4X2֠mY:zYSdG=y EK ?:)uRewm9_Ẁn]tq]_c%{!:T5[7 ܋c %Jl:O+:_3Gp -26wUd_;G?|>?8H-v AQw | ʫ;W`i {~  !\nNk-XyЇ'q(1e0͢^q!z| A4nC<a:μt@m=q}x"!ik?GVev V@,-64A@o=u!$j"Dբ=A'_Qri@Vf]('@ pw-Q[(iE(#c%\(&4Ѫl7  ։Ĥ;yघKe@qp.WC 9}%G zi:UuԬ2I3'Z NI aw8#I!h2۪J1é|y^Q8YG R;S6r(E@q~jO4P1dǹ+c"랠:~FRtD*3zyMBOL=.]-eo`Nmh9a/N6#q@fiuʆBr79>NBJЏf*`6!C I.Z~nz3-v7Ou}~)gdBeGT"2Ppx s4P v`uf5.9R!e.DW]@NWлF;ݩm'1 w6L~>h0oq&1zQa0B2IEJAc]:n( mEׯ^eqJS3ed)U"$XayOrm`!msOˁ}_ECY> eJL5TJXR(E9$HX<8#"g z#gd, &:[,T )ˑ HpLUqscNiq'*@JJBeHTHp峴 Q k8YrbrȲ!1Ʃ )s]h @a--[=rFUf_P @/|+Ue|ş'ijyјekJ?˖_I=/ WK͏Q+ҪH~o6|7;e&RE!!2Ks\Ljb ) ` 8ӌ]5r=yX<];ɻbNf`P!PO♓ofu(I/j^ܡ}AA'|;o(1^+ /wU77;fE{1`( HE!qtf_ *o+m ڣs[@8rN60Ög[&Ww2F⢫o wm;SY6 4$@gμAΖ-f!̫}~\\;Y  h9R3Z 4 XiMZANΝ,բ eʋ d.2ͤTVߘT(n_iW:mx7?MSMT(%767Բm\Ùh qK>3gGk0i^n|S6O.S7ٓoMoReh)5ѡZ7 /.SM[FV~c?Ѻ!/|z)N~`,=Z7_ \3X(Nκ'H:Z3OQ9eՅ 5rh݈CnLg4n=ƸNKg?Ѻ!/|z)w;ZX DgH0 xG>Jjs!ZjdX ȋ; 'Xgᣟ0 !VBO&(OHI/ODO+S.O v%./OO+Ar|y~O+A).O[ F?aB b+'~J`_iuZu7xJ?CT5#0pH͇^v^\!C͒R@ mmAiۇw->E`!lh8Pd5L$ʍ뫭INR vr65TUHXPE$C1S!&)?^|?W>~и> TѬdOo ߾{ bQ))o.{޾ylN.|ٝ{"XP3&?T,jq7S)??wrͺw֐t7VfA 'w7+*siR=ԣ XD;j)WZ!tY뇷~B˚hGyӖLoeBHn7o{`KenZfZ[]U=FDZuFwꮴe|&orv{Wk{,J%=Y} C/ l" `TPVo~׀E]]=w}CQLхUƥ~X|)V `^oy?r!̵`{*cPzMDj۽P\ fkFY?>/ 'x ZG[馯lXCϗ#DRA%Rat Wb QNh_;m0x{ołe>00.fz4N_dqZ3<1zQͿkrOOw/d%>jO wrpדxnt~ZVjN{i$9L4" Qbe$!(>HxߜʞPY s1iՍ G[Z=*\[hdHZ "z'+8 p J~~3? 8D;8a}jHK۴pù}f›,#7L`Ӏoö# +Ǥ%XGa@BmɣK6ixѠxKb=e@^gd78Vċ-4ީՏ#WH`%AuK:v!&qp]1 q[6^yXO )p$\Nu>*s9HnsKR-vj1M≾)h~،|%A=kYa1>Id,TD'yڍj ~Ne,jRpT[Eq$*bԛI8ITqgs}ˌ]͈T[zlЩC*Klm׆4홈.x؞ۂYZcC/6L[n|W4N{ʲg`NęLUbs&\%0@9Ey>5Z =c ñyJKC"73[_`r7/ D&\XBX& p#I $O2%91V4MbjGg!Fb?؈@ب]$dd"r, 8fV! '(MYLxHOQ0E^kbxD@cnu^T(ϥv |PH1 ^Cɭwި[b G/l#Y_١~^l%0WYl~8$R\ LjˤN|iT!=`IЪ(StqblIyU#g/:r޼mQB`#Z8A(gP!B[ urY9_E[`J`%Rl^=Jjn9&l &^ܟlv &t zR]-Gks# (B׻QaϽ}C{sB5W垷sW\bt{Q-9 +: qyXerkNw7߭&XJIGYo(}ܠ[xWea%%zL^!m)"Fq8kf1:y`Ֆ4LuK疨b|m⃯$ÉߠZ,9*ԠkuëGtkLu#A( c$>PR)bIJU10ŞD@{bQ\'b.}b (\O}<ˣOblJ2]&S]b4VIbA eyɴn#k0.HzIvi~:>DPf}Z;7QQ,Zvnݹ˧rdrivj^ j=Ȁ* y]EC#zK(Tyjbz]7sܳ[ F3q=Pgzn8A>s>Ub s;ԃH<*ٳVƹn=?+UP8kƅ&grmsO!f!ߕ." ~[@б2FeE7[&3ߚVyE/bnGPOz_]IȪ+̀NyjO{HHZAL :魒 t adT`']u+)8/eMۑp)F9w|11wxuoTKօpmmST Z+A,IG]8tkY s!o;_eyݣs?3NQT^U:v}1NŲ(! Y$ M]hDShj8KJyFr&D)KI cJ(ei&(eQTeiEɐ,u*(x>VX?a1sQ[$N`?LPz/?3*~}0KeSD("I2v~g2<8Kǹ49q2J4N8) R4q:ߴrF )gIZDJ0 тLiZ&L CíY Z[{jݹ0e[ay]:~([5A x32eILL*TlQ=mQ}lD_n&`"F$[D~x3pXn6;t*d ƛߎ} ^Oat',Y,߭ ?,zM;ϓl 0N$hFkxX@~՞|OJ1@j5yo5)Ԥ0m&9\$Yrq^l4Nf'VNSvJQ ȾLrTGӉRr<5 U\E'8!Lg@*X䔦:)88Dve˽5-Iz8K&ZX -KQ%$'%,E "^2Ōg[D2 Q攗`SLkg.h)%%h)Y0˳Qȕ2* ), VvإHC˒KQAiVHcHn)Ց,%L`2eL"-up²A@p4GʳR= uBK{s.VfR~Uw "\=}xM [c.U {_ C53`h`6_ϗWUVZr%;[^'@{[ZZIO{4)2&] _RI4U7hI@C*Ub`&9 #{σQ!4VY4aE' +X4 M%S@#F{D# \j@< D w7/:m:֍ DC A; DTnP1~"Tv,ZZsNR{P֚8 |Rn"Pn_hW^ㅭdLFDܶ?@rfۆ!B]%{£7 Gp$n֞.D_\Y}y0W7m+r k)UX lsy'mII@4x.[["%L[9dϚzr,FGԓ.l0`:bxP:t8 DB!5BvOtqJzIJbdawA^Ql4ͽ"`O8a#@`Rذl3y uub`A.MKQjkr䟢h y&eS2=&)e!x7_ lL.mT&ӝv-}z7bKօp)On zb`NwGKy4:dޅw;n^6i{MaDdꋁŻ:Sp@,[MM1*\9 Q xVl{)N0VYIiO At h =XXJI-{/weҨe]Ժ936L0nl'غQ/._V7ҔGo>|+BX\]^Z=}%XJ\XN'7UPoPC;TrY ]S#ou< ZKokr֠*# Lk]}GPw;:8a\w38YzO9+ĝ?GoLضܙh/_cIx7MNrj!%G ,e=kz "3|",O7*7GBwU(Rfy)SQ1f8Y*ẐwU(DӏMsYP%V:  . ꄷ!V!~aRk8_weT Fx簾F[ K^1ȯhsaدFL%z*~Dwը_j9" svƊlMǹ]8EQe'Fbv?iX%hby~ cQƢ.E]-jX1,qS , TD,",UK9A'-G3֧1T% /:> bc߬"RKN>*McNsL&R:D  j$-H%e i LupBTh e S* cA"M3$0Y$ _4Wy*|ςy,nd:lO~, womQ}} 1acjHKC{WAbu@|_fp֪5m&buI/ҎICR_=F4:i>"jB qKz{kKRq,[x=b^1Qb`zg!:B6jG% jJF<'L )}  NJV v y!ޫZ+JQKqsM{.y]YрB ĉZO1mK;G{ 63=_ؔ)0vJKLG4(*eR(ӂ&y4(HJdJ`JvOnM E JXk\DJSAT! A"FK)' O.) Rp/07 -MgsY,&kKŝF.X,mVzfj6.j6Ɋ^sm| !:=9MZC53P|Ȩl wDw0c1ek+.(՛^C.A7#] eBpKag"(aR?(0S\ɄR%ɽWkyH+ 4նS]g_5d{]s>5Dn=={*e}8SIB'6-H,'G?4|9DDH\D-2no&QyuX]~S-5LoG_ξXɩPfՀc͒ݺH)QEԻƂ?hyAǂۍ0t^lA@u#TBT]RK_.k56RR:ں siA w ?ܠb츻#TE^u[o69(58=ҲKjk)fNR{n\X{;Rs ~"IjYe&zeں` >ǎL<~d9NLM?h(΢͗朗Xds6J7ӛpO#]ys^;65zXT2jtZksZIW0*Fc~`VPje=CgbYSع# y&cS7ݔ`!x7_ lL.m= ֢obݺn^6U[1C'y7Fd}w1̾wB^8xz߻}wqһua!/D۔Txx$b2;"wSk܎,DL9v$ Z][s#+,-hTݍ+u|.9yIJŢ%jN*= 6HjsPz+@@޺MV{Cz~߼|n[h7>5ִmMяOՎtW;ëjG?Nr^|A;0~$q~$MXH6k-j$RҬb3uh1$+f62/+8}zED_|{%p\Fݶj-lW&!,(Pۻ YЬ*t} WTAiޘcS>VtЎzKmҴDZZ®JU-Ԋ9 w}l 6М%,|LiMV`]ro ,Ɨ̖e,(sy>xjAtm%]x쭛6 w'}dTjۙJ#>69WK>ހq]G3R q[|K8Jɑ}z%N ۫K9Xt~09-׵-ge{O[ 8]Ʃx/ۊc{#Nhrw\@kX{giY["o3*q< ޴b޽scwG^{Zz^!Y=n ;`J31 HxS:F<.E'r[iA5d__?~d?I3ק3-x=X5_t.AǹKqKjvQhڙ2`#h[>u ΃ P%0ҷ^$Z{$'h4*GӫhxfyOhiqOs/4q.F}?_twB!ݭ+ZI5-^1pGpj)~x32mH fg$tє3 ƋG+'dk>NI&]]7W}F憟ӓ?Qy6BX]4;]ϒ'mgqtrq74V~%e!Q8FT?v.T%ZS֞d 5jD/ip,iœZQj0tlR&}~1n+z]]Ϳ VZO/W!e|7EB) l0r8:<} *^ki^+v%! 7]3{c2W,]_UgZlZ1bޅr5]*;++Ps6D&D4F"²r:. P5 oCU6k|#:kI+ B$JhD Z/%}b)D!'WF8nBuXU۳zuS Wx]T4^ <%/ Z-ot XUp$<> 3$lhV6t߳iXQkݕ1(,EF1]I& @ H\p.@2-+uzze%.O dqFryFPAՌ )Ygjf&[II8Ef=PqɠHX:ʭ&@a"Qw]gUsI,%B%ˈE)VpIʉ>%>gsN*G/fC.3XNj"yv>W 1Ù'Dc`&]_=-jrNks$_}m$pV_^ zC؛8_̏oӓph5N;mz8 3w:'3 ynIW"E Gx!yb'e-w4^A51OGS-^JO B I)\;}dGӥ#C9%Nqf:k'R8Ki~j?ՉyձQRx6$ u"^#Is28K̭De4%ҽ`(7a[ CEwFonGyCqdLGZ7ޓ5# x(JhJ$Dj@i%2?~?^x@aOi 櫇ZӃY'W1DVȕnd#vc1ΑphȭiTa\2(2+ 3ā1|-6ߧ~~B} 3]S@դLH!-+Ctȹ)ƒ@9Z8mDdO:D y1û Um!N LM1|&0.HH5C b-G} f9}z$BvVcp#UD2FΌp*9 F$NJV3&0lP$[Rp"Ē /|OU:(cB $C +;_WC 5B CXXFiF%;| a%``(Y \`JPD#{ZhBDh RZ6Ic`: C<~Kuh?PAN9Hpw!bέw{g83_1'L ϗ1RFNOchrr+IeD[2ҼOVƆ iÁK>sP eyL۬3SDk=/4k6zB3䦚,j)-4fH64(,[h9͈i,4i@ҊK,wemH02;K w][11;kGfljR1@%G@U,Jėĕ_fvCSmSgCHp8x0dA-yĚ=+i\(bcƳqKfn< 0j@jsmfmbV13zvUԬ⪈ٴ̬ ꪨ(ڄ-oa7fڒjW l,E5f&TXʬATռ]1׻}AQ-} ٗ?Cm]sILs$ľYq[Kfl_ t`1Gb ƞU$l̔S΋JjL !#H\n~œ4_W}yofV܊{ދ47E=f6"/?/WY *&/~%5/ @ 3;F qo9nq^]& B B+dZE+9: -\sec*n'SݥV\8i6R7Vץ-6znT{cV:O r6{s2rͪnlg!20Kqm9e=KY˻fҳ/bh$Ѿ)T?:T'>_6_?/ seڧ?zx8=w6S=F?w* DM4ʦM&wtbng1 ̻ůu=[ y&eS:ټq:Ͽ)GAN{9uN K} n,GuJ$a, 51U3m|šN'pT(~Ci@w2 BXbb&Cy'(47pTK0fM?CJFᐲh772.I.`x/oV~ə1@+X?]WsGҮS։-jacS2J7L^7)z3}}suD9ۢ꾏SI4NEœD=H(zO"(<${ PؼMW%^!ape8d IW<Wsd@tS vXXOO1,R1g T81HNkHw GP6=ƫ$] ]ΧvTEkGEp##xJg3J;NS)zUqZ/QHůb/蒪W~ TcWDz.cYB_K8VSL)VE/>5f8/ x{Yj{;~l]<8/nݗ7'gzvvy->8[}ҹƃl4>yy<9ZAWZF42HN3H`-Y13+!73Ƙ:6(g\ ĄPtf~T8/*LQ Q0zYI 40A \j EVP3t AyEN1) 2//!fT+Z($u2D%𩸼R.ܞ$ IJjR%Sf0>71nLmE(J#wkHaY یK 2D*m/eR()XQ[*%AuwaKaﯛveMTQ|*L%Q*pS) 42ՆPa9#E-6PgQXGqݝ#޿y@I&1bO ڟc]FϥKBx!~8Oɩ ueJV"xIV6} ڷ&{PB$3w#>,ƀjDT@K(UIg㌄_uhB+5"-~\hX\qݥ6˰Dfg3!>_g{XԘm#4Ə逅N2(),䅛hM97 bc:mxCγMڷw_zzM4Ʀ(c['uSAS:' G 2:'AZCnb&1lc4&cl#Ė2ҟ90uދv'R`]S/!^1CVwqAA Qc$ѣ]t]0N#vɄӃv1R 4Wx zZszp싞Br\*RiڢHQieժb Ck);50P钇=>>meKD <+)J9eER?i P m-i^hN *XPQb^ExQTҨB2+zGtJ5g\E6Lob- }P~ )b@ʀI1,@2ےӕ?CJ&-|\"{ףXbI\0AD%8k%0ۣ(('BSq!O(8 =Qȍm0}}suD9ۢ꾏ȁ8p䛞8ylǠ憙@$mt-H&MLT]d #[anmfclo2Q.%<9[Ӹ{#,[,LX{vqq}g6r:#DVp!3pA rgS¹2c[ʩV41ʉʧ-m/Eա`8ut>RD!j! pܷ8?}:~vaA:.@6ۧ@~ѳE}S% _aR34g_D)1H=Kܗ w$EX1 AJ6`uCp>x@TӖ6 =I } {}iJ-b+%SQojUuA zؗZՈZQ*_:Bx9é.9/'@[]?D8R|p8b5V AXJ^1q&]- @bq\1W2~%>{+ & I0+iW(/^0D= .E .i*/+ 2{,LuY*!2M\L3A QZResGk9Bjq0$l4 cI|@Wht*`7gv),䅛hMibkkdv|].ŝDE"WT$o<}g֎[QR7:0Zʸ -Y!9I Y+4f4 CS&t1eRӌx %fRSRjGs($&lg7R;Ji/5q4 q 8'kzPX)G\YH?f}VcIo^o``K0ް糋[&O3=Ͷ9i`1B5>x1bad(":/5":dži "={G7ObѨN%\wpϛA#./O˹ĴS\\~y}'n쭶$/m>_M맿,? v}7W46RrBhsO5緓7fu(U?M4.2 |pϛ;*_zW*,a{|~ֱgzvvy-읒g SԭAjp.z/X}1ua cT1}1Th&]jҽ5匬:I"+0͉=3 j~mC+=mj?W~r7rjLXH YwklKgŃv dybhc:/;WwW}ސ/?{G؍IrA#(n-!8u"X jL|ۈ U1Q0f5a/zqyׄiKܽS$)8UD"rF굈:*dz3,,JIt28׈"$ՕEsR ZLsBCIӷu`iN%m%P_4Oȼ0$#[wƏ/3B@ˀ׋n:7ПEf:0fühl:#Q 0}|=r%:gVI9@.'޽]T iz;&yzqw48z7&c@L(L ~tt.1ۅUc/#&É 7,*%i92Ӎcgtsl*Ï닻(&ZnNV8Q&t1s 9F-8;ȬsgnZc (H:9zD|ʳ s Xor=^NMhV JC?e졦NȾL}&kRߪ>ii(CKc;Fo޵Ƒ$Bb/"taXxs(-If&/fuVUE"6XEDfdf֐*Fϭ;̺w"V=n=XC )iDZ^vrDD93]>޳߫ƍ&O=9 XS\ dV*냛#oUz˙YVI<Ƕ\^t<* FD*0ghv,@tL#{qH-ήh bc[E%CDD <'7G*r9%PizNG3Q<k#QGT x[.ZS.Z U@gjńhx +ȌV:+㘡A,Ē"F90A[U|f{aIgυ+!q׺@ E[0-I‡L-r+V椤,U ִ<4MVG#ȅ" 1&6-MF& h.B"4G(q#гeJi~O:I5.ٽeL2v92,[\XKV+ckvT,Q$H.x$1-g9vTOK-i1ud^zǓ >9yyZ2H dj) 4!31VdXj+&~ = œf3jS ѭM-|6VϩQ;g-BzH՗OrK_~oM[PMg{g̰#0ۛ>'hk>'}9aJ{xł걌I=/\FQb'`aMi[@ӢIJ|p~J!ۼW޿]XeLgݟԕ1|KFKzsj ~yn?m9/j|oůowY:NUbiYEfMLPwтEͿx7Hֳ41uTg2p)3f;ۆ'j1(1gN@:Ӎn/Z: a!/D+l(uc8r~DZAڧ5UQ 5T 5zhK #n*꒛K' AMi=* ]*{bb 119ܙh'G J0Cw~,=:?xyes;Fr! h[[a'cq6~yJXNZ/v#Y&K:-KN 6jA/cOZYUO~A˫Tj%6MQ3hƇ;CN)-kejInY6Jp鰑1ƂMV{E~qPĽv*M[J=~BП| tۓ(q:<}лU1F^jiy)mfU^o؇-4ZK~FNP%X bD%\([m`\蹋=K ,c&O H&[U7S`!bh:ouJ%p9XJ+]f {lcJgU``;c+vl!6!F!qA[|+p.ItvpqSfPhT0tVp0Abd1ӝ۫^bv۵N(̮P fh0]" 9Y)J$9#" E.(W#Zoן{_C%- kf(-a@j!]^}!Yw٭*Tpj0ΎF m;2 Aniɟ=E}:ݚ9o8 &6rD (J2Ovu,v(Tv/o>}}-s93w6W]\٫F1W^TҘE?%E9*[V?g\Tts:;.| @'j ]9ɤOPdBA۔q[ nW>q]l]_T58vwFKH ZW5uYeX]&B">O7vsx=&\Ji y?/; scskOYפy#5׍SB`!n-7dF 1.bQZ!mڈ9ӪeAҭIY_o^X4 Tgy'G .eqdOey@4{q0PrN*Adx ֎mmy3WE-mF _E?Э81s2Y*3C ^vONzB5xTA6d '(ᬉV5']>b`hss8ZR3`1rhPۉCrt(*j ֝4rbPB@V>ypiclI#o y&zM1e#(>KldӝXG㩁[Z"9vVvt)6C#ʷhZCA2ۆe@KvPd7MGxsYNTyHblr"ztRGn,ErI 6b (DldY uǴ%=UT:TAeB B*5&^ׂb: BC1'")Nʭ[ܹ3Vm: SfVb_"r%َefz'e\KGmĹZhу)}j21sd||ƱsdUR=YqiѠiǦh;6XK˶bEV؊.GŶ~!7~~[H9Ď>seHgs+޷u~% Y?qL0,1ĸQIΓpee`a;tn w5j^*|X(meN7 1Q0&Srk3:ئZ8l!N1hfY ,x$ZW-O2{+vudQ18-dVSXz8bB9P 8WTHͻu"aS&;`dq!+əOBz3P*Mu Nce--ҋ9^[S]}*UערAr~ 5 !zĤZ~>N6sEim32Ӽ@J̼uhGT G vNXڪ%?߭z,tD]?\d0P F=;~{/eKZ8d4M5|lPЀϛRrBPqWɝ =h(Cm 9|*~GlG#,"w"pO5QF߳"wfq Z?~KO׊F_yyEzx w$;Q}AߓuPv_/yR3=T Uу Fc3c7[|]m(v)^F)>?oe2W)[ElQ٢.mh,*c. cYk&p\6|x- kTLG-:I!ۂ99+ЏaZr(|*˒qM -{%0QM3PZc"uPs^4c fouuE1 Gxo[&MN5 6`a>hTדӵVW9s*:*76 S8/[rmPAhH-]dDMYr[gښ_QmO6ĩ:{vR<%5SI&[AmQ(TcQ@fOtB #~!eMCDB(݀0pW.c]!e¿čWAf&K3V>}G?/|1H(_Uy s(syIEoc+ˈ+]eܰ%*%1ZwnolJFڔ@}d{%Ս(cYY{|T; CDZٹij(1`5!?^A5]>"KGWdcQ.h״F9D7Ou5L.zXjOuCB+ _V9KPě?MGmsk(7:?5۫W3嗫/z.IN ;EAQa&ɝaBARG^m%DiǮ ӊDׂxo-)5ZYҩlc U*U$+Ep7<4D}0ԋ,mvihJɫ2t"Қ,):x)7wƦ„nܐyJyS &Lg9u>˵WL !-} r㛹P趺Wm<*At:jrw($&n f|# 9"4Zԉũmj.'uQTF)Qx ɧB[DS.Jcgl]#4!C~weQb hMJ0nP{ߛ2;VrW$ADٱ)kk:Bpx"Aw"CƯb27KA!,SgQ1quY.D-+rpVs J,Qm,؍-b}oFO] upG9[VO^50ʾ.WhQS9%(l%=Cw'(/Lr~;!o9M9L#w[b䜓l ^=tV mD^P[޵<#qIZg"?&3;9ucwhh օ?G 5!i]Ո!x(l;Z*]O|1ǼbixNXǵ4^SZ*t9GTF lX0 !&q0sSL-(7`%֒2JȵF:l)V(asiNG/D]n?o_3@dUO9tI@zz䞖>/JxD?>=QB)~\l̗|w)<ɪgiC/?|O>YgЇO;3_b  `lD-GWaЅ)$/g|nǺF%l *<ˑ[Gx,me񃧔U/{%^L(0C#[JH+Td =h/ '\¤Q0nsT 3-ةדCd $W tt*c\HqlS-ύϩ0:py.sy-R2GzšGt:3[Is\(ͼ=DTR+Lr@9)AH & qihINO -+*L9D{H HƥuxasmFй7X >WyDr(DDuz޷L K: `IbqH1"srҮDB3nj1 GA JYδgTQQrU8ZW$MdYuV춁܂R!"U[Th[ADh3 B*QnǺF%8CCF[.{o[P]Y`YmJ=bF7`~=9r&/d1L!2Lt!)4' Moބey6THEuᔱ9@[OțFi, |9ˑ\Ql[ZZ=VX4d:WBL 4^9%Fa%r Lp=H`Bq5Ī#پuM2ʒ^@/Kb1j*3@0hW+ 7ӺZ *İ0E!s@bM"?+52XLdEygMdBWvɾqD"eM\BrX[7#.)EЏpjHĴuFNr#nP؅J` xxiT"x#e&5BІ1PyUff6J;jTd`^EmA6s,v2o28[PM^mU,O-<<.;[A;}C"j[X9(8qBq bgLSOD*%QE aݙ=FH iκ^F-?]|,eCT#EhrgY [0K0`[.*-ƭ*-F=bڊ/C_^z$=V=ƣ+.]bTH D!MÚ&jS]jh_mPOuR !p}o#%m;EPe8aCqyj0ŧG:rEazmABh_PRH;o` `T2˜d˗5G}fsXJl+U AR $ u,W8۲1]מ-fPiy:P~hT Ӂ LbǺ| QP_A1pˡ+~U#SK!Ҥ i6W*/XT\,i:0 GGW)Dгx6Yo\F|\V&ग़[}|8Ez]%5_ޞJ˅p`EjKÀ>ݣyLB :% pE/.*2z>{mOp}T19)V@SFM++KYRX*'/[wnI6%'MV=osLA$Ȼo*ݟwK+Zwn 60P|ߞJ)- :L [g v<QTNuh~hSe{_!_}ΚX߿=BFzj7exX!^+p`q=V{ _׸b` \]Ȳr5ws]_1\%0Ez {w6ⲫ _0<)b9^LQG #@3Tw@kH%XޣŌh=ky7UmЋBCzvk9o9[K aT Sq Bh^Y.xAt3{zXc17JV==I06s;F;(T_6?ҞfƸbĊ">ꊆ"m4i--VK62;If~+0bxwQ@ 4S͸J⃿(1dȫQ:jęwmCF"%c'`AC_#sJɽ+Fp,Ud˧{b\*=Q&6I*UqόkIɠ46ܮ'چ}Oj.NKZpZgEjB N'v @-Guރz]?αV:ąWk]ԔOwy5Ra b>21Ԗe PBLTiayC0CAiӫ5n+nteA+ *O*sDv\*m.ϭHc,Ѓ !Ph!+H"~SU Q%Sh,5FT0 qBS@()SL@,I}ɤ\meug4P7dAm5PGP^WNҔTGԡcQVGWsˏ_o^Yx? %ғԐpwTq}2+֖ܰHqF҆"G&G^P;)!WߣEU$HIKRraE] Q^rR־*4LKMxdPXSɸ:cƱ,ggicNm+x>r[whmDӵ2jehx#KoE-DvT(krp#~jKݣ? 5w ҬWN 4X!O$1; Uً>lx㝢a0V*ztݒڰwnI6%7rUWS11>xmbEܟwKhwkB޹&TeclwU#zT bL'&k$`rz[EHֆs=ܦbUl9om-w*ɻ+>7-]:5a~2{ > ŝUP=_A#lfZ=]T3Ip۶z^Cگun7&δ(DѶA>gƛ@ H9M[ 8V}GBeiRB`XH74y(–lj7ai:zMo˟G.庰&/3aY)$X!f Mi_~5]~22U+q˃+k1-1x+|Ʉdtd1$k>/s5#&\x/ r* -4O f v J%%N`!]Kg5`$kD脰 q=/-$RFk7#ϊD?!8fL:<2*0SW8"Ε; GD5E•a׼_%b ,I0 pbVO2؉|T?>=?EJZ|~M|\y`^6qˏ~0Sx!Up \ _~8wbO>YgЇO;3_Ɍ  `l;8 1/_} SR E5x<,F_piTrpw-dgk_ _W62^j?=z3Y~'vI"bR|YEʬ.0~:#&'E[ W3k 3 x%(0qEkGÁF3u\R{vaJ%Št= "q׍ȼ rE ;:ejN0VX.Xe*/iv43ݏ[؅eJ@3T!*ؓgj6_a^WxRW|pIb"SZ׷~=% lC,Q 鞞g͔y"Lq꣖XiڍF$BlB:|~˪rT2K+ԞRBSf ͸K.RW\V`NkHU;En𯄍`zX1|ʪ]sy+g ($jX?f5@KVzeX )NT>گ"~%9XQ,jg;O-bqjd0Zxۋooƨ=oVJ-ָjϽ7 p WbͱBesY߬l^o8?~~"[Z_EM4raY`mwʹʹʹʹ*g*qg bXW)Dr>j1% {vt q]؄tyk߭=n$`Xf;2TW2wWeHl>G?'!)E1vҐpp?C L`-!rrY7Sx~qmnWl Es,O(ya<`7K\@(@4ZksNhqDГBK++9C#csgi%1\co2<'>Bx!oZJɝvF'w\Q;DL80B:G^B .%?nmnf '^OU. ta9ۤK`U\1/RZyzd,$ LM0P ^tO>X)4ăKG\-) Qj5R` aϙN<1ǃ4@bx#Abt bJM`59rxAXn cK鮋 J9{^U`5p96 ` !`%f\sZ 5ef.b2LX{!??+&n)&{Ϯڣ|v-gGIΓ9>|UJσ(+L6E2=;E 틶 ef9ҕ3}OH#ӟ(dB'Nuc"#,HpJ _S+B6AarHGn +Jm$;xk_ZQ8JЕ*ݟynIޗ@gY7_IC[)]F_5CvE1CGçЮ>{9s |sUOiCz r)b, gZ\?P fژumV6kMY[w ?UvxV' ))xuy뺀X/FOu]4*FjJcHGGWd cJXP<-O1/9v-MQQ㇎[8kPHbCKoys"U oi ;Fޱ܅8M(8dy)"|wݖV%ur3=Wǚ2vOI(\ǑsYw`7ò=~Y2GeL)LzCZ 귙wF 1"A" "=3d00Fq"OFPg+(IJ` sj|^ҧlailW9,x0TyƒVtW;I"gs !R&UjAh[5*dZ˝儀V3)+'SA,8:pI@(֣5B|C!{hD$"@9Y$rt(cnKKQՒ\9QU"HIZGݧ`FS@$aAZtt o hʼn8dߓ(toݷ-;Sī)q&h}-,L,&-CXYY> PEdEtdz!m6/{z'%crj qMUD'Bj 6& DHr:^ B(ޞ砭nPWBb5݉Χx4'rZ!gdL 3AZ2wĜcuw7$!= ҘMaYUhu@Rs[wE4" )Œ5l3=wCP.ewG4ѽ%ۘ`ޙN=lb`Љ6f,]9wm߾;{^_~8XI]=+N8Q#_ ">&q8&~*S&8%饒m{^QT.NrVO[׭:QIM9g+ܡOT՜p +r /o"v \D6r9xX*l'዗^ݗ;YZF  (k;?"|%|r\*o$88m4˵, lpVYC{7WЫvy}XvTQ!'Rc窓׃bL=tglSi#1#U-z1ggr0П=f8z) 3n~] Z}v`S_(Rb5!ݏ;Zڳ`P u`{^Q&ʧL%XX%|Fϳku? T $Lbql,#G̡P̢2!ݦc mޡJ${Dx'9p8T`I$fM'WhVLQraY7"΄q&3Eu&ܾW QemYGeixǐqIg'^juUԧԆ^43*ڶ/ZMlV:.\r㧛.D 4n?QY,kh.*kJI T9GYe |m;WNIG[ZJ"Fy]? Y)PI aKrPLd:&<!͛K)jC9O,,Ju=@- \x1h>3ƯV_my7`gw~mw{usJx[Ak98r!ORؑ/wW~]|f:T\Y`|O ]#~SQey oOhMa'z7V);wKŠ餶ŻsLѼ[:)tKև|&zM1֣]E2f=s#E >qNtL>S$<=QPeCh yIM7OLd"&թD%VLq(tJu5Xnji.MYj: Ƚt , Z35M Y` )%΃~.ug6lŮUmщ룚6]HtO N[A_P1q5C#Yb.7t+xD6\ura/t>~*W=YHB{Jdm07,9= >;4`9Qo)RUa983}ӗm>: .Q>πG@k60s\dUn-;J-81^~48ݠd ^__}R1/[Jm-U U$" +@VCLSp sY_=qOruvn1 W$%8v@UAjvGZk̄v&@4i]&(%JJLjܞwդ/^Kѹo|^|q,{%tdU2z)'>*.K*u{%bZl>$wq{P8(5% σF k.PJS{n$;`h4ߡc!~xpm%* \k2ꫣ^z5ƘశUBpX yn 4FԔn멽Vw_^ҙF"HYAJL3S2g02IE$"'dzV/4X7T $W/D0mR]Dsse%JФ̹2:;mJaH*4  A $E$,1SVsL s+0q% \At(!A.(ih8([hA ;iT2YRNTHdIX( RT4ک7,Z陴Wnhߥ}s2a]22AF_rwqjʫdEQ3 NFCB1**nӍyG^!p^!JTu(D!j~ySM3٫08^($nKEn,G5"X[i'נ|_n$J7bv>V(PUam_@B<]L13fEPUR\\G0`p S.ysɡ-l@;M쵼ˎ ԰0bߵ\N Ϙ4 &[ Zg\L|ׂg)x.d5_gDvv擫?zrhɆ*#?m[īb cyJ| $8/r S)U(^0rPCSVjAس`F-5mQ8*LIx'X%&Tsߖ>ڶѶ-}j4C2, ,E H%XHdsH%*͝K}Pҟ}޹њ2$ N/K")%eYZ%Z""+sMyB%zaT UVutJ.B|-e҉ ;4pyL+zE EPr< t%f4Mi:۠b4ӄrM5$ /rse8Ijy`+GbkHC36(Ys3>3FG,U LBQP&+8ܔ1"Mw+`RjE!=h/xbrS-LG^# ͔LI,L.fFY}̵:{- _cٓU~M`s,[!BsVڭwֱƬdKD !!{9/zw7Z[#d*1j@Fŗ4[37ŗݚ̿_'ki':slI~Mn7[xZ}C?Q2/e#XqaInCк%/C u XjB19} -:m֚)K%aMۍсcevA W)I_5ɧQLG !m;Yf j&j;)Ryf\Yh'I)::SKY3oyђTI)zhSR] @BT ̜ Rj~y85|#ziOĈ{*^PI\ \-DQfE0Q)Z܀G PP &&P2c"I 8L8-0- e80fX0rU? TA{6PwQllET"Fo0&~J9٘Λdg͐'ExxZ%?--S]zآ!D3TR.?Ѷ]I1ݐwɣҵ)SWK'~/nˇ6l_Mrֺ{E2xa[Z7Nn>iyDi} /#!KV7wX$?/6s' h 5mcScք-9:Ҏϫu,as)̨LAV_~-0 Ϙ.R]GTܨpNmk tq*PA=IUTI t/ȽcPF47s/'_bơ #2z63Vhx֪O˫MK]ɰm-zz[%Intb)7#b'Pꓝ٬xp2kDIdpʹbڧJix+E#U4; ~7Lf*ܞ/^NKon| IŗyZJJ;^st݉6|/NRWW 6yA,ͅAk)XI+$7*9Ҝ,OBD~xaEK7ݭ`[I.-nOj%1>_m_Av\Zo\6n_0*a/*$J$^[T(k+TJ׌ #uI(S[75.T|9LXIZ7'r-Uն컧qhvQiF#$H(+7tRsLfei9*(2) /K{X-2Pl$O,%MD2ՅMyǷ~yAr _Qpj:u6sݝἱ]N rӻ7Oٶ,]H6<<ɮ.̿O]r{cS9>ny/>ck}obUvP yMO╪SuG7lY8}6l8h>z&zҦ˕0$ $09=w8RT+fVlVmw۠1"5#Q pjb=~X+f6\ٿ [*KZ k3>.w*1xWwl_~z}]q<+PptJ^ʋ2Ee%kX*\4lvaEYȩ5^jKUre&­ƕ{;Tvt&`?v\L|0E|Z=/>{3TтlxC}u4eX~]`luM=:#Vz_0KV(Ey t ]zYMJ '>hY V+|ކ4Ꮫ+$'@̀ torIb47?pZĀ3]=2Ԝ\qpF{BC+ ERdE?:ʍ\0F\ᤞ0`o@m+GʾYkᮐ`8/6y::P[ ৊tI(v *^8 /O_t`BPTR]ۯ".3 !c30a+9:B\X#^ac):ȶ&tFvڠ4stb9$:M'3s?~36X}ކ=G$$W]PgLrNN@cN^h|;+>ˊ Њ N,G<^]r]r[ J>ދT`:E; #ro(*FW0GkxFu8oOYӽz-TkuʙD$UݚzQ_^YڃֿNJ/zL"|I$b> |ܝɟ^*K\$^-ޠ&L\ tft!XH"u#J=U9rer.O^<rJqNl8V9fZ=Fv"uI2.U_P*Eۻ29Ex}LzX *W[G3aIЕCaj˹þT[M<^^P*y;,eYjc6Y jT];tl/*aQy~O# QC ϸ~`Z|޻;O2\G}8@wԄp$qY9|>kp5<W4t?(m<5Xc@G7F$5ȎFGFG氢 yg̾%wg&Cic~7&Y^v*=l9ٸ$-6O(JKRn忟IK;!GS.;5 fDŽd &..E񗝢|% 9 M8[q2JQ0=?{B*WKj)e"ʀ0"0Qx3JÞ{~@FD+'42XVXMc"ᠽ$^!A3sN<—wH]Aϱa!RAf2%3SsbFI*yDO 4BskJ#vp<~ S\TqMap"x P3B,|j :p7!P go|F%(BW L@L c:. +&/0P3Sh4AFfhA "㌬ ҿWBGtz5-az}U"ՑVfLS(ሂbR"j87W0IL% F|?=^T327%;wK$6 ha˽?ck4̩W/f>}%)ǘG3SzRf%7SNV"V8o\|^_SEa<o}Mi}ùBV/'X~zм.c,03q_ V2q[kΕ w_R:j1!XX X$CR;p$)'ݿoVe.7SZ%$h;cDRѷowI;x4 y^y(#@0B$y(9~FQ{쌟!w78}Qfυc/n*QR")2 eiI" xGr^Ќ +Q|1ً>UFe Ť?/sJIE)RDA7~Y-dBgz9Bsg< ?H!hr}^8ՈV;yyk_%TPH6GFJ5tg䠺APWf6 ¸M0A@^_Sw_B;2EcQVG'[;?&᣻s3%SUsg*|^hL&.$}]~7&y=~vw#bW?V?_^OEۅ,/JXH::߅W:+AJ~d7> n0I׿̉\.^Th{W YRZ| 4䅫hNU5Zoݺ ֭)mu;]c^ymV\քp=Z0RG!PsC0w_πZ7a&?pZ%~gA{u9v1轟_:Tx_'3z' e`yیmrs>h_UrӇEhGv0j!p y10'>lC';zic\ @炭v_F NڅCqֳDK.TXT K̕rg߿%&,IOHlNzb4HuzOcd>ҀZp)R`= -)^'j53 eK193Bnp-sO1Z-*Kh9EC EL$p !\zjvX۳Ķ b 0$Zf6C1rawА%_޾=;jp~t-m&~-[|6d^nn|\d1 FY˚~4T7auOO Z<-\nn"Q>N`18a{PQ9wk"**{XGRq:PQ ?ר eVá6\xqP>rd@ɯU)g[n Q-Rޓ3kBUv;8Kj8 nTB& OATN`qFx!8X.6(l?Օw~kOvd^% Lo(=]׆ۀwvl y:>[^$J\oaߚaU :hr-9kJT# )> ̈́:T^[?8h|/EǾT8-4 X WVgnN?XJJ[ٯk10Vrrz nv{vemM6:|m1<_/$?5m8b)3} ^(49x4XZ"|(,{(-B(À5LP%.T8>'b \16f?oO" ƳШnƋGs*%fx2RMStN GlFGei(&Ȥ!e!|:5 B s < 1wToD& *R6k}oQ\5R;J|WȯMXWk|/UJp(@i5k~$uK5,:(c]|PQ\ Xy `+&K;wdN_|ڜN ?-(e` f?_g4ݤ&l7ՙm7ԇc.8bg=~nF8 Z ?R>9y~{{"qҫKz[008a=㩇# Jy?qP_>'pq=Ќ<}~;?_]_~&&qyGU_$E˺襐 _n$F\3ÀFF$8ClZ8@8ԱT&PA(6z &iH \!!ڨ"äօc:)^ߺL( 3Amc180F{eLZ{dI sy$ʀxkN03mVI"AD wmN, e.h*_AT=VpRƘL5fLRO5;u _4`cxЄS9/FuQULrp< mc Bh4+6mkDr)Fr(xF\h>&ަ^֤osT؞]X@N;rNeK5Ms*cfY9_(elG<\2m9PXbfM镅y7wxfw1%EHED+ɛ<)`yeWS;3A96%%NKf03[ghCq)"sH8+-s"Dy FH˜gʚɍ_ae_\7}ێ];Xe6:p845M)xI90TDD&D1v694ok֬0b++$8 dRe*A*tig8c19hK{J`Э-Fi3Z5JR6)4CKV [%խĉ!W$! [L)%t`W0Ԃ)k6E>rq_p ({F_!<>3^2ͤLOθY3E,ͤFsDԾi_8jn_h]v;4Ns6ŸV|o;NCӈR裃74M8€TV[ĩ@cypf1FDoufN* 5MWz{6+ϑ~fiOK {1sg~#f )Fv=8RDuU~;ъl kݹMӆ8nP<]N\NJ1*t$8t"TϿa{%_F  pΣ,/W*tfhRB8"u @P^Xά *_3C2Ѹ`]gFvtPMD YL{L`d1X ^e-I A5z&i+ "NHH 9q)x"+-5ZA3Iw5mN+%NLbhH!<(m,8ꬑjo#IH* ]Z+KMosC}c`Ts83^9|Lg/p |s]}'v?5Ơx^;YZTO..h5/o0|0wy5{34׭DjvhK*NUq_QYivyEjbCf]PiO_l,LIdC(KoKNHDJ#EoEz"Ŕ0 R>$B>E):SbK>f`J qc 1 gS?mRM S Y%* #hSaXN+%JJ\2 $2{Pqo5`|6rE;AG! 3m0H {8 {Pۊ)hN0'PH42X )0e$27$Ͻr]%P@-<-?OK!sNiii \X~ﯗ\|DA=Xlcu~}͎ 0#Ut~eZ9տ6|Vݜn=^\^8\)l/rbN)H>):s`";m$wFKn75ɗU.doܖ3.N=@58;_fˉ|/kf뛶%}?mKZN Z :UDp> QKn"T V/(K}P/j Ht@05cTM㍗&b,sG+q¸AO*kRbl'I$ g2rD]0)sh0”+Y IJ2mzQH5VݨSq]*D]4H033 -ŏH0M2J2&0AwxqQ*t.LI0Čm2 8XVsSn喍+NmEķPP_~Ma7?75)f S\cg٤Gv0غW8Lv"Wǃ~UQLk(c\>}+ VoE1.F=5j/\R篝2vL  mpmv%퀶y8 JO+zXCd9mJXGF΋F%p,1hrʦܰ-.m%i `Nw%ђIJW76Oa(E9fIXeaYgo>2Qz 9)=j^2d(W(?ҭ9d*-iWEmG/\Wzr /!`"Ҋ*:Vڰ5lTJ^-|,L"Ǐ|H{*7<[㲓ߕ5AG wyʏ"Iذ; NΕxm+RU>QF;E2#k`ٝl)#5F3,sJ>(ioe$?(kU\D zڜfƦwW3?V.-bB /V![lUhZj?T$:i7 CidTܫd*,d 'kqk\.&_$d}dO;cзG&~,wNqtz **۝M|]Ju- %+"7+X:#W*Y%DgKpIqM)^Bƒ$^5nɱw9z>8R8zyn Z_G6BNMPX ([)+fWʕdfk<=2Z9>ZL-ޝ 4\F S4o4[v>j۫ ],7Á6r{f@ fl|iT~KVl1w/:kfs2Jj+ P%cdDjS!-fR0f2B(=`Zyu1㺴X-[yCސV}iKP=xۊJnv:iVٞ!rZY^wZFZiPQ>Roza~`mFVRm`N°WM w*ApJiqB+N J1Gh…dIsK,x8s9 JQD`E}ml,ޢW;fn{Y21ThYFWq&8  , ijy e9&4ەaT1cE%GX ZPRWVTI$@D夢[dh ehL>/~_N޺/Ԇ\|R ~P^ˀ*0S$3&.bAS+<OpHFl# 1JF2KuW=M\`w=Qs45V^O}=^6fBn]&F2y/'f)HFoJ-@HYTO\Q7`lѧٔa0OQwנ[l5Qhzur[g0=+DW -KRvkwa^b+t/\r6JG3/gc8"h,,퓶V#ڦ {(r r4RŬA"%ő|zb#YLp$އ#p$9~J2Lg)`^>}p-&c v`/8-l)?'(# پ&RPBc<`͠ ոeIβQko5٠o5t妕Vo^ t޻S]NaWFՅ)!kkn{kknRw1+i=uAݫsXAz/sһ,T~x񳿹 K0S[N}rIEw<_M1tÖNUs/D{.y}xai=OeʏqPR6|&˦đm4w_ѻA}Gv@S#,nCXwn}l8:7s ޭ JL;xW"\̻o(nCXwn6e mL[>,j?'9*l=esi]=8ɥnv~~5sWջ7礌yN0-'AijW*ٱ)mv!0D [3m,ԇDE4H== }p_{ ƜaFQ|103:HI ]d-[T᫹^$'`6:Cf^ztHPD'I$A̗0f+d8AkUي[*& I%lb/D@%# {:J$JP#8J9 .PͽrZ2zNb_βת`/?]??\[4\lV מ|^L~s+̲7ftgL':\,n/ρ)SN/rbfj'.!GU7r7^B6}كv[)Tp&T@82B%ZaajP9 icfCP,;j`G4lݴ: 7B}}d&DecPTYEUቷ`!װd)nR2b"c<(<~ J~m1J@:< ]'u,Y)IšfJT u^!{NIx8N^rREZb-XdvRhbKFݱ2 Rkտ1C,tV5q4kR("ID%j3X1yK,eԬ{LFQe\769ɦP\FwJU>) Q9 597HjIcURz/[dsKʫT&ȧeV1qH h-q&XP={-TJP oYitYY%{W٨ЦQ9S {F6:WSLn?DwwN-f}q]f<0Q=nGf_xM4^ˇ[Qk1fAJ~]2 8{Tvm_?OEZyCB v)a]B<$!B!סYإsO7Z6Ly VC̃CӍs5VSb@fǤjtYmˊ%N___ۯݹG}[}$tt*ʕs܏7& lH!G*2zQsMCg8%V2_~=__O~Oo.W`û$Q'nM2eN&-_MrJw WKCpɰS>I^cMk;cVnW׮6fF qM&h U._JsQS[J\έAס6ʼۄMו"^0Il%G߷HsfC>M{39vʝ=yOxsvk_6ZvZT3 V´ ~fS 6k=eW=vHw|$1ڴërK,۹ջ濭zvw7Z v4S2M% fIP!~AaHMQޙZS*!y Y Nyzt5ם|04nH02z'3X9"G縘@t@iı)/ Qtsoc3W1y˕rKfh$^ÈlpZktgNk6" ? MmMBmDk7+mu GϫS*@};+n}% -4A fM8~1w \Ȍf֩wW㇛8s"a;㜳$C}e^ Rj#ap9cXdי/jz|HHJ$%TYEˊti5bS凋hB%rE(bA|woOSw X^}>5Oަg⮱WG@m[6{H݇L;-Y 4|U~pɒ kL9ݺVIDp\&!jk> / ,l(bՔWӭ]7FN E'lj<ȧz@Ӗ}\5ګa0D<F%HԤՕ;;_MLeaX+r* A PR.rfϗ 9-En"H;DFJ޲B8^5($uS/|$g tuF I! M^mD i2k$xPqR;.'^}QR6+IDi#BHY8bJD/yiŴ'5FJ!dhXҡtEEIIs2* ~P#5Ԩu0Ρ[*)rYV8%2 N)yr-j`p!K\j&c5C9 &" 2'" 582ѧ/QՊX gE:Q6XUHd%yD9TL&u2+g 3HΨ\I0FS: B /XT,kgt pԼq2ǨkRJs2K PϵZ 4X .@6)QS4@Z6!81(=T^/CgB ŝNPaMD5Tvcʆm8(W0yIZ:r[3|ZSl] qWyvF9 ֝o^˷V?HuJEۦ3`]3WLlo4zw|NZ-)Z#iOnu?T6͂<uʭ\:;_ \]YNDb 0ghw ?p] 2w:Aǡ-1U1@HF2ZTA*PJB)xEH(TҏNDmmY$߰mDo0)BR 4%YAE "S6U;*NQ ?yf"j1' ;%:+ef*!)eەԛwfDl(`TodaS,ZlD46NI,Mg5S*A!lJZwfܸušu [Vz> [1nG-TlMMFS [sҐAuJ1ywpr89OHVnravjgrs/j{FaI~{ϑi]Y7Y)9ٍhg֝^\t׷Ṃ߯<6w$z 3@֨1n01u/COq>@݋PY+ DW fS|za0 0-#A ۀT`$78@^%".-imR t?>#V,룶z~s/{VDJܻ;?>73&On9>4sRǝpwj'L).Ez/)>*K7vzӞt%zaiIF6IgϽnbJdMsv5 ef<%ãǖXnAZOdbQI%ؔɠ2eu?J_ ih4uY{CE JΫi+Hwqۺ4 $EI89}KyyIȶVJDl~6@C^ޡZ\HKmH|UR@y ڰwcHGp c{N@ٵMze,4Sj;X愜D|=Đ(m%?'C W\RB78=]TX,;׽g!{Rds*G|w1} q}OqyN Kvl>%.Y[A@Diٍ~2+? E'*24XK-D\qZ*X8\E%Sq݊kw ,ZL酋AR(b`Hz3mv ;4tA uhN?_ W"U)/lc CC =Qzx;|(OZ5VYG'tl]0|6}z(Ϯ5$Y)+1.G[&=|Z(ֵPyK=q6o5M|6+_Ik$EyNCMG .eњB7<J;C_R`(Dcq4JZm~HnjΦBXd.NWR4Nkݺd|%v5 vAMNJ1k~?:4՚?jQQҢZg1ժCimݲKKgƒZAnQJtY&ufvZ!ײVN';(b睊Tm%q׍wV״=JˊmVfbpۄjNy^1jŰ](g 3'=FlBZ\o_R)2`yy^҃\U4(UywT7jb^kh&vI:TsrRnjONwQ{Vi kOUj^ɬf%!sBT R}-(}NR?URW8J]3V)B꿍!ޑ?_\#xY_* +'0l\^w| -y))~_]!ݬ j)ď{h.m˧>!;9L#ØKܫBCƒ7\U(ƀocN+Y>IcR`3#0xH!(Y_0'[_hاP9QF_ߏs^M9>oq2م\2ѝb(.9LjS R$AG3'5vt0ֱ{>Rgƒ͢/p!̉A7*Ҕx 8{@l:|LΈW^?s2|ϟ[ gSRv(`M̫&)KM:N ) k}?B71WftjQ%uJzBOWiVa/j &:\ yQQ(YiY +=Oj占U~TʻCD&/Qk3,6C̗ڲbrN2 ER)1yRTܛyD2ף6ϳ1@jɇUbBDNqqQzJ "c{5 Ǭ~?[jgh3gcDk˘G3*<+<{a%hʾz!3QcBu0jN0`ƥ}vL/f8K e^ pVI#FXU:̷V߷f>vf7=v|>>VBo7\s كN='mڰRKz@z])Q$6ݘ 4N*)`e^5FM~Ӆ[]NA7V0W2vK9%N~daj0ɇ^%-~CtLsVG<((I9bS/xQ <7H*StDK,O6ǮS\8) \厪E)/~Y$Caਂ9)CH/}bxtL{N~>/W9&6xôK.t΍ֵu; `aR['9sFmpӮ =_OHMK3e"OVh{2"392f;|#stO9?Ertb4 7Lcbk>*3:n\u8cջau: 9bpVḩkqB4$o붦~9^wb)g;[P+N ?ߕJs="-);wgH9VKP<ΞI9D+ ڬ]}$\MP݆S'1TmS;ȭQÞXTU[T^x!kbJ~8ttIu!Ƽl["HaB_'7%r<›ގ)Ov1mǧRvb,p`2oXL|we7NH ެawmC<05m0ynGI|\gl TMo#Wnox$YJ#;vp5#zC@#tS 5C:ʖkeoM $gtq<8_2K P1 !V'`L,#{Z?nƅeުaYXHTљsi3hl"v#|I6 /1|W0;OwEJigzȑ_g#K;X`Lp0lr,{$ٙSlv}QK댭X_UdU.aFVlvne2CHe@@@HmO.Am3U)IO fRmfw#1Y[hdcJ a$W>mfwQ}N5'LuY*#Td"RP 9swW,6x0L,u  \i`tbL٠z^,tHUzMo^:ʵB85pؑ@Le6ţ^Ga1#\,߽&L,r`d*:avQ}N5UMQ<Y30y\:) ;OGʁw2)O ba  jjB*PKc jk9)ymORYZGRA݋?EE)D-}熘)db<&6_ .l("dfV/߈U mU'C}0n5mlk8lCÜ1 #Mjy-6Fp\'gxGGXSya0A _:/Hu)¨8r8:5Ԫ:;E12sDPT4 }qRR0Fe97i.kV6DfF6~3gw{~yM&coazR2U9K%؉W{7jpDr,UEldM)!R"P o ۯJr+,#z`bt֪lUfeixXZ,SgүX]jni56\oܞXkÑl WS]v)7{H+\"q[rrz5v ,״U`WkޝjuLyiskbv*lV#9\XH)AΫZ) V=ə黻/-`s7]%Bq |ndN_YFw x3~'+I_Yd @C]D.P|LݵX/NjZX3ĻpXZڇV %hwm]Kû&)uXs)7hN8aE#ܪ ?,v"۾kAT떀 ,;f BN!u@w%º[ûֵS| )[J۾߉u{au%]~Rۇx;G;){/;|P^h՗W\ ҩ` ֱELJ{2u%YnQ ן&]%O-{YoOp ,De.ur|LUSW7'&He>&)̃^m97#iP$5؝ r #*+s갗TC^'m77bg%:kyEIQZ]YpI1QN7H/C0gv"EHn-$4Wj{,iӶdh\er7_Lo쥪lKٽzrXhpZw.FJo7ZQ]e91*cܬ#o14  K%=cwYg%t,.z{n`~4~Bh %bGr`3| 7S-BSBیs" rRඊ73 +OkI/@*dE[<:ɨI<RCzUc 5 jO-oc)6DTjrqݿ͸ΕeXWg!@xbfW័2.Kl1pAfsAş ݃{pvu~`9%>ITRrom&sSNYbGBSI30ޠӻ-rѫ7 01T;mm7q~d/\k;ծʵ/X5l"iLN1~dReY.#rc9c1[b3EZ2pNE֤:dKͮQX8@Ȍ{֦fV/8|?'Vٛ~WSx2v ]+tuBy+_ޏa1)KU _CR\zyd">ˆx KN-xAD C#_AVA|xlN5 l˷{ _1f7{c)©x9 {wK0Xab " 'ns<$C"0[>~%{ YIה8)h$SC5n~?6`ܦ.1+aoHAjVrsjqZK ^wv0_"*r~8wpX9SQ2W#x=yײAzƑ!B q^V|[؄Uvo?ϿBq܀P4%-p 1 '^OgwE˝ż`:ǫ@ :7ܬfhr(.,g,~/9 w?XqNJ(:۳h䏾El"t*9da{PQ`\Ԍ2pR"a~He:,ڝ<~ $° 8w))͘NV?<򿜡kFs̉2$MRJJEs'|qX FS1yOlrCl~9K%RT 5gQpDnԊ );Fn!g'}˰dDHEZ jÆ{O6y H)|p&c@$o2RES]1*b$RG DZIFJ'ID)X2CI@L* Frr^1p)x~k*\붚hdPn JJi|M4DҬ$Jlt(5T`Jbk*1rs<1$ZjDˀDT e4)TE4cACF@x=dZ vT#lK9(&fXk"pM!g(l|"56EH0pQS$%'Tyq=HȰ6xF<5k78|BjJ\%´!)Q 6r [8ܼG+i3FA|GǨ$, 2D@k.)zaAQ-ȚB ¶osFV~zRYearx~󵎲HԯBj`6{D۴P=/<&g]fj@m'Ҕ hCzb/e@* T6QWnVj\٢ HH)A3M@?{ƍ\nHp>lVj7W%嚧3EpۺR)>fJ9hFhСёB4xwv.XOW*|lKWՅW~J2@}fp[@XVX۰ʇsx8{s}}XR41,jM^Ehoupu3G%>^w3/9Mh̹^EaX3hKrm,uq$'R Cs5 7ض\eӀH&[.yVW(/7^5͋e;'ykWW>NWy@K~x7xS7/gŲ+5-Wьz4\cr]Ur';Re!E|8u&y,kKt/;P";n׉,[y.SGq>n l\.ѭU!_Ϳcet !ETdE8Y1y*{jٹv(f^ U%Ga@Si*7?̋e-=j=tG#_}7qk2Y'=Вļ~Z!Q?98U" @58p:Gq_G$@s73P^Ԕ=LAJ3*Xhw^OTJoFf-Tq: byO$LUm8(BIuOHx?ڷjm^ j]k67yQc`/}lpӥ0:H_΃uɦ&J au|ϣ5`N&D0[Zr.6Dp;4o ez6o=fUXIYg=Ghri'iτM+f8lm{fpw;havGQ<,FlD SM~\j.E |wļyӲ6ba{MY=a78F5l8qgLXWRk:j7}m`S v'j7w,k7Cx+wQ/b2It@ l\.ѭvxƟ2W^>Ed􋴧M%kKtqM07~n!<䕻hc┚==*90a:0{yrmAb({J'R3A/euGY/jT=}1gzRy;)=b~WYo rGy,'zk=P7Q8a;8' Vmn7y9$9ؠK7BU{K6Z2sb` #@<9Tb{F OjXL+ȹ"+лq`ce#߈Wkhe죵Y2au[Hβ6VwVcذk:ydΪbr@TeƟgSw(T*NGn n1;,q6[5lӼ5,Nɪ؅:7(yMVCwsak-!Ҝe4RH˼ s(AVh=KEVYD@ 'E 2$ !k (9|[6 Oz7n.ZPƪR7ofee4sH^挔 V(C9.!J뛤>[2 ]b%O>IT@\AIJ[ǾߴyQi^9$Y|~)f?m/Hޮb6QN{HFuV'_ٗ\Y7"69nj㩞go!#xJh0sZWsOorMöܵAeBs#w9N!Z[%, @zq~CrB0Aj}D=~G̶HtbrS9])fz^#ܑp6 ߹La8;Z!FOoiF2oı՚ ϋdH9-,/xB2 aDGL2P_bL >JLjt89Z"YX0Wm|&k~ ÙU2Z0CJhGSZ$ʠkbSMjkZ|ryz_NMce,K>c!(U8 !!iDv JJa :RL@ :RD9Q#{N\PD+>PDH ۾ p3JH'^>$+T+$F 9SCHBN  \U>|/ hѪ&cRjf)9#1"Jͻ)\Ӧx'BoombO2dwzw3̾bTlQ"Cc&~gK%|=I5y GӋ,'G]ɣ*[SQ/ywɻP҃ʉ Ow 3YXd!_Gl iClm$d:/_kMI씃Veo"t!ج{=|?ɝS%8*nQkl#VUkB܂Gl?_ >A2ڝ7r(dK^faCe`Y]7ԗXm%Vm>`zJ,zp>a/yTpp]V4=C@hH/)q z+(j)2B~YDnC|I}!I-#+&GY,`}wv_jŒ bgImAHrW{ ] H|C1cpY<2+riIԟ2 iJDc58` @(I1JiSf>K[_|ՔJ uӃ|]?A^Uÿ~ھA ~|nm|NHRIٲzEQusz >Vbڠ~=xw9BͲ=(Ӈϔ$Ya?>6B)$&O'FHlTK+v}S|#"Yew?YtkC/  ;B|J}2?I;/Er!FᥧݗZK Ru,EM>X2!{eA"+xL,RHTi,q sDrIf*+LRF"I*/O(3*#=_*/z8_˗k6xF-7Q4b; W$Mdݷp})Pmd1F,$睵/B3@`{]EL|@QWmLBm2l~w.neiav6Bh:Ųa̓u˰=,Fa.bL$/ql6m\DyNZsI-{[+) C.l9SNA1-VJs}M H.#+bdz=+ASboM+B,Y1]6ӹ7Mt8emv:Ͱf߼Qt-ܩzX_+ZMbe%/.N ~Dp)R;M0A @t.HP8X؝fDb;$TDTvAl;w}~5 S 2/@>,MᜂD&çޞ;{ynז; EHlyrXڠyR IĤ68}AxVKh=U @<+Rzhj_:T- FAޗh$ƠxYuFzYp,HoF 8G12YP/ 50X,Z@D$14nL[P2b:jt]" :8r?XU7zFeԛ_Uʕld#wYs} ~b"$%a MWg1lx{4QV&T3Q(.S!$A '" &`eCR '% y1&t<ˀIFi B0)QAGP 8 ,rʁ,sJS"%>&U 69/%q&Fc\ZT33eJ}GIGfKƥ-եz3LFB#0[K *%L 5&Oȅ9Q]n֎tUh. )ʲ"O *'ZΓ>Kf*TTU[/lW7W(V*[oUNG@.]LVw]j`cl?(U4&pVY*@sLZ'+j:t\m ~7oXIdT\u. R9<1ڱ^ᝌ'8|;]𩞫"vjdB%GԭJ_6J\MxzZk-@*QbymB«+jxkVK1;3cͤ Tgj Ud8ohԆy1L҆ >q{szjbPV,Zq(V6uCT,ɿ}Z-?2h |5}orx_=ô6_[oWZ@=ZN*C0]/IqC[vOa>y A3|]dT!󘟋L>멖K//}չX]ZWb˱W h92}EejHp!L$R(|E9!"i@g$Uc Q;";m@fֳq qޛGg='`HBۉYړ`em/TM+N)GwGn}F&5S1GR=ۚ*ɹ_{r-%Uyؚ#9+Q_X?_x:B+TObv`o0>H=xE^[ٝA*#bbB vgvIP-MaDw ^uYi\joԭڨuo8KZ:ڃ[ڃ k=R8{k祜q6# {Rbnj@+bgݗZ`:mc]䀓B ٚcV7fIՔ@ھc8=-3yƟv%)$:cg)8+&3ZoT\EBNP,Ʈ2y rbmw̼η 9gȬ+@̆ dH Ά)S*8-ĉ(d~̀IOt5*PՑގ~G?*$tyzG#yR;VE|/UNQ cu3V+AU_/~?K7^޵$EЗ]~3`% li)*A0IYdJjQ1 "WU]/#\~>U9j|[]K+&jA1_ .B_赘[;vPU&y~`{[S֑y}Nհ>t֍cmB!:;O:A(Q.JG\>i0r4|}ITO6\F\nxzz>z`;L//6>i %de< &aHb)k[eiˏ}=22Q(<&B!Ґ  1?q0P*a.EDýjfo?*0P/^ _TIO7׽c&'HXx2]_]UvUOyCNfVK\8.CJ0ׅT[#;a .:f%cs]d1XV`X Q nhl;9}sjrkj59}7>49}cK%`MU5ً\խXL!D\QّҵzoQfduY0+3Ƭkz]۫׵o)bBܺݘ@sqnU˛FՃM, _QХ3ޝ svEwΊzu[өBS{:狠ϑF-CRrOu:yfb@Dy\As2ՂP@Uݻݥlӓzv#oKҹܫzyG0YRW,њG?-!8R:/2>w6vK Agd]ۛ3H>"=5ewU,HJ=)z^/t^黿ߓIf$9($䕋hbW]}GnR.X)z5uf׍mjI;6$[Ieޔm;S\oG6`pdK$hzLq‚Vu6n@7rFNr]f~A]dQUݑ?]U?wiAy63gAklMS:{'tH<\ǹ n5| 12=ؑ8܆ kBVW~Ue햾[UY;]z!3+.(fm}kcN8ɮnb2j9SQmSaݹmj+ RZK鿠^mVyZXM Q4U:]} d_vt{F|7;y yes.:cQ.s*DcMz= ggU} kcNOI: Mª =}QXxB`F"o$΋6 $T bJ#Ŕac:"4R@#sF8[٧lSҖS8#+ gi;I lMr\c*x,AB$̊1"b,Һ~u!zga!6믮*wߦհYg'I1Lwr&otQFԒLr#J}*!!b S9x{ej;[x|u>==OOTdcWWeVpK~z|lR4_|Kn><$v1D &Gy%"/VFpMEBr:3R 1ooB(pTpJ 6LEj> )AIxq?Zցb |7~ρXh+`Ij7$O*P%q!֒%ӏ&B4+ئ,ʢ17ԃ\9 hATqOa<`FØE y(F hByaG-&w\+b3`e:2Nm*Z,Z$5Y!$SrֿL70 1X[Vb!R*!`* GoB'cP2$үTC,$wJtJ8qXk GY"fu[7.'["aυ"SBl|Q8&, E/xdO4-anwӗYċvl8 +ŧ|84aY0Ŧ>M ۛ[ m;>2UUxOhm"_)A񵛧=\(FQ7BVtKtv( :)=c[6ʌ;BV3b eP>HM!6ixSQ-&kԋLPuöytۚ0%D{;ֱ,Ɋ/VŦVgSSyu%_鸮|ՂLU fT0}U kTymKC;q0շTsџp)MJ1VI )MJS9 R(jd_Lw|x@:ARԶMmhElb1\RcY5( )K.eq:K J#YKkF.vdJv-dbr"q|J)dՋ0iL42Hx$}Kq+W.rmDHoV}} =6bYٽ05ыw #S^=+*njhoJ{|]o8kVYL5<$?@P;/Ajչrwr5Ѽ-vwr]dew?nH_̇;>{rΉ ;`B&3_%QalaEUbL0E,U#h Lɐ3eݝv(0C,:EZ2z.T 0PP0YasJơDT *e45bRFY*QV/gTfyIAE 1*1,* R !L ,Yfhk'ѢR,PJUfypRqJ DiulcbT$`C r'ia{֨*mEhsn[Rq i־e=Ȓe ϘXM.+,&OYL}BTXQBia+dKYh)*QЎ:e)D\!Th$.|n _R 'DX 142 o6+l9D.;5)P1E(yxy џU2 8݄Β뭴vC!UuﳛN ϶x> ⮝8%;s\TNDzhJ(L[RNd&3EJtX;,\0*,\V%gϧuk5x}@4UsbƟdM{^xL:C1PFjζe9 D]7LDK*;qaO!n)( $h{$vϐ\N~@|TCдt׵&^B2QeroXe[WVeoaFFnƅR 7s{^w_apݱ%jX;x3|kvAJL>`{,}~]Gu9^f-)2`!(X*C(vZ)nꙔLQ1 G{ e,8Q"Pɰ&<𐚐 D1|/X$AH%)eئ%&ס%F\l{-]3 3e.U ,YvNw?KtҿNgQ;'c?Yzbb4wFI8f|N֘Ki)saW})gם'N' YWI[L914շ?,Qy ,X4#ၝ" SEd-7]PTlQ$[d^w$[z_ň*?VRZnFۻ|޺u)n'R$J'6Vڵ.uڟsg{Ę QC]2 K3D iŹh Z^ءAHuncgp8՜5ٜ,Ax ]yzw\.%NTFmb։.TG k9Ǫ|MsÊ<-"qnK$E۝.BJ""9|"a4EѬ"wBg gt)-mE-K:asCv1`,!H[ɕN( %SN }@sѕOӌ8']K8[\,n:w$D߂BPqVO<-;\UY_(;SJ>Q{t̋]*%F1vO(um cpZG<Є式QUW(͜7`ݳfsl5GXg7Zd?FLW/X&3(cG #sN穝Y͂|>pDRDGYt ^4LAg ш"5gDM?-DeF EiOhwWn^ $0Adj.| &wAwYbVuu؟.{Oُ`K7}yz2k8gtp`UA~KAjnV*SL[Yw~f_&q1[{ۻc;C&i,MVȜu/&[nٵpM`wR&݂)|H зLn:mh26fq} 7gG!qAv^URg5:ǃ/[h _itZPpf.{ <oY֫z;O/# դR՟fC|y~"W0jכXƏ&·?flII_{a ~ݿgdh-c 2`0{4 9ǟe:h6xe(w>Dp ]X|bثkiN?eb<~1&OOA\~g_ I ?nߙ4.f Bs}jBJζ`0Y6G, c|ЯfP>=F$_D3w¦ีܟ,eE0ū#X=,&搜p D"ϗHTzRh ʵk34ѷ=afb!FǾO/e҆R;vӻgCsptBf+sJNFQ"p{ݜ R(*@xFNՓl|m/~il_-JC^϶]UPRս_Ja,plh@Ih (Z@ J"*-ERLD_Mͯk<Ğ<]P;7ch A" JD,! 9e1H :6 Zj$6HG3uDդ!߹V)gTWMIXSe1BhͺUOnuhwPB*9#*W/~{ 6K](d%!wyb׷O =UuW2WƧU`J #N%\>#ePcLL#cdd+FPH1 Y:%\&:i3 pcG(ҍG<`RRH`*# AdlRa `6`\fCdV޲Y.Ř!E$͆16+b1eR:Xn+ :Tks0?e#AY9b!Qw')7m^:5z*= TgxJ qv2X-剖,F"ZxrZXF`/.|"=c|$Ԏw| k5>ȃ NeT ּ!\<_DԞةبG9{ aLL-Ԡ8  \"tz[4W/NzW|++$\I Ba"t# *@1Z!G4 %K`LaW=ړkѻX=ۼzSJ8aW$PXpekxS#d(JiAIlX\9j,7r+ CZã"xGd/"56|&yHuV6vg-U*d33BƓ :kQݠ̕R6]k}SmGNKwNA7bj 0;"y/,Vd#,[;ԛdjRBu[ A/V$GO@Z_?xE5j>.HK{ @QCH"k!o\Etj&#MúbU/qNjk0 y*S=-yuK FuRbݦEkiА7 :3JcqGϵyCk1u ,y`U"9rfwuHtXͶi1HIXӡD&j~7@\Q)ViQJKmrrJb 2 멚ˋnXUid4Y]۟7N^ Lձm\ZᴑY"r6"i-Y(”BهLEBI# @j`K N΍S aI`LڮC#ݟ\w -]t_7J`.O˟+!i7=ރJM43Xp׮)fhc{\%×ʁ]\ɬbƻJ< ^-9f,S 9-+R/~->.J;xǓm,DqkZTJ(ddRǵO1*Ħ黀qV&&=0#ɹ_-]_BQ.iYs:ĠL /ؼ;:$qYwR #.( + F)-;Nr^1` ÉTΘB"_hx)74'A]NxaG["[A4r\ mV4{(n/'W{ ^TNJ ֒q/NDYF7xF/ $m/Z3iYj AjOAܨlsBc6Sz_ XȈ߬B ,ڟKϞ ANzSOp&8R#Č X`*:xݢ]0duU].$B6X#/ (9$\zc0!ġxoS7]7NWO#.Htӣ>Ou(GFz_49 H x`H׺`-Τ/E}OBnEnRKaN3xwLw7xuс vW!G&0 (U`inf-ɗEtn&/蒪xɍ47yP1*H8)F\!]x -˒ 9En9wހ8'9A)ݠ!ceG2 7ށswe=q/סlыq>Kw1VeJD{gwWwwקQ&3S%zGy~+OT~6<û!D6WnyU.͕} k U{էa|?Yo^[ {ɜ!MYHYpQ")^]Wq+on6)+.@ˢ;+75Bjp4֖KKWY掰@N(e;x8oǰ~P1ȋ*Sd$}K+Ghg/ 1pU%3)5V0цK$Ih^DV+#3V'^@Q֒I89ƑЏUF[k\MfIJsCGhS4_`AF6u6'TEP*q1Je4 "b;ո+d(Op+߅-%j! j2JYM\ADd!d\0AFL/Wp"CC*J,bғK-X2dd<3;A :y eS0@̸F+4ͬ!"_nY3czKZ.Kȹ$`PCŠڞ_ʹ؞_be_R~nh)3-ܑ{>Ȥ ToOdX\^j*h o'3gYRUBL ?$ "F~yߘ]W;'k;y?\va|GqUƥ\PM~}4*!2E#SL2Ȕ"S r, M)dik<<Ԟ|q镃) ֽ9~WhΫE} _|N[at4@n+n~.^?TLg{C$;Ʃy+ im?7tqݏ[S w9spNta3C3=t޾3I鉇ļe!1/]yOC?Kxݤ QtʽvtzѥNF/镝$Jm4^9%38Z$32:,t3Ѝnw=s&l$>2B(`9[s[G<H?t`-5P<9R[w^oF"qW~HǨ>, CrT rsI5%͡1v*1]S{:{eJ91v y*SBs]H:떊AꤎźMwѬ[z@euCC޸&Ծ;=m4떊AꤎźMѬ[zFuCC޸6tJFn@0[YsbZEbi d@8M*:,D^\>[zEkLAQywX:gh4EJUZҼt~C]VL _..•FDFwgb%Zyc#WҐ˄S˩Uuaqhb Jt۸2B+t$ѭO2 KPڻȅ"@-cvVKkRFx?s[ %@{̪{R2{JErb)'=W*mp$T >vG~x1ҹRNގ AV<*| X";@^۹~R&#KH)J}H(]e8}˧()8۔T_nVW(8ʳp?a~wҼ*pRm=I5>?M~9X̋/ː`?S. }л@zGd߷+rЈgEec|dkțE=BLvjf]!S~x:#l>|ؗ@TkR n!k< -B@KI5ZzZ"e(fb{);VKM#aybϚ*Pk>|%_qY E~ rj%t| 0"ڞvj݌6<H'ߤ+=EĨUAaT{.ob!ʲBsQ ȉsEkiJfdI:Bn5 U`bT4OP([yV.*xY:GI R0b͉Ce"%`(H6QR 6H~#ZUk?Zk֖ ~.u' $0.-qFe$ Bty`Gnp3Bh]eNU.j1~"R E+/`Rv%+ (b>wvoϞy wkz4^i_v-\nO߰ߝuȁoNCx߶6 yh|7Eĩ)Bc|2AVM; gdQvU:0^{]/sv~>Ǡ^~ z7WҠфbh)zBIbH.B t$Vxf %59+^=pS%SbaoHtnea]Pk9Ƀ7#2L &k):1qXqsn+yт\mNR2VPu:AVv~ܞϛWQнt )2vn-#c iZY5|j@&iAKJW1 rޟ5(rya24H &+mv@;05xkfu^j/qA^xx(Q XgJ񾺞%ϝݮ['(wz7Dr{2DU~AdLyXH ޳$R@k`,-éCvUIR;~ 7BvQ3LC"("tj#/[5g=^!6Sa})DYlDQ#z tԲSDZ9mDJ*՜ 3: >:EF1!f+uc7Dz2uqT-Zz&̻ GBEzG/3̕-A;QaD r5qwx7o}~ٓIH|nVct튕ȕjk_e./cׇ?\:\|'_/Vo[~qL ->T~ڟ]d<'GT3c,ύ.&H_L`2Y {Bwnp~o]h6C N> 8Kat7hH@$I׊I ~:Y}ss7jeOx8Qr@i0T . zUNAɢߟju;Љt N}ځ^̓'z72K<6=^Z.'@iP"VA&( TaFjkEX9"&yb}"gzlX\xkzj|H tbl\|16A+CV?.t2n[w}avںBOp3v&C3T퓻ru ޓqM))Rwnٻ6nl]+@/P[l a8C&nX+mh"y)QQ$5s~w ߼-_\|H!(<~wyx/ >fl$ \jhvJ%ż x=gQ.VޫH(*Vz#K-yRf ,-́Z5Ft1uqgoa1@΅)TlLXUhD-JQ8p"Nl_62̋WKs77lx?5X#Yz&ȌعSAO'\'g8\(xi52&^P$nI|58L?\EQn8C65\1ޠ1ڒt>}]h :QHSvXpZ:?\vD{u@99CHnP4wX:Ԫν(Ͽ6E5 "E8cgD󷅶ʼnPbQdnPXU6,'b-œNp W .O\4d1T |Q}E5j]/JS-Rrc%tTORnw:U]ҽ@|ThM1>zorkuJVTN;bVN^\ \ڰDlJ1\Reb:Y<;%7n\ڰDO)*9d}h:@en<׃)Ժ f/c>rBV= ʉ*J(>1灲\~c"?V\,A[6G-ctڣK}f?< bY,,9=ߛ;jSvZ| GwYqF zWO_9緍Z Rg+Mo+,sVģ6eȴg]{wM_~Hg|3=O7YDPs!Awv ӽv7mp(fG=v Dۿv{HQufaRtU{>BuHB1i)~V]7$8մ[U5,\bWNLqu[enS-WSAH^| G g\"hc$`W̉ԚH .ҩ0L`;/<:G t61uXƂ $6»gtFƴ-5*3 _e Xa*,09fX45 f}!G $-ٺ#3_8Y::n4Y's\-=8 RΦQeʜ*S2ž2XȔJ$k]-61ٸw VL30g}/_7i']ѭ,][F7EXk*jdkчVRC(,]eHcFȒR!VKõA,gše@ i (@r#{f@ 8)j$UB0^\Zcj-`C d t\ie %(Rf0q1 U %mg>J4!F>|+IZ{47dJaIƟ?&=}zvoJF!,Lѿo|7"Iw砊Jj.* F8M1jT 485 snf9?JS}ݤZ+=K+VcN t4GA˗ݰoJ.GG`Y8AET}>ߑMڧZYO2Uo T SN=X"\3wuLjHCrzI+­VZA^p$ XOї-B+!7C+OcEpY? ОwMI"PIbuqLX]0!{~^fGF(B̊8\ŭM9Ys zY}jMe聳Lob:P/,&AW=Ӈ[wy1air'>d3o3o3oYG(7)H0&.j}c-I܂ "y[5dQb6ܓ[/FcoYqZm)_ްhrJ-v ͌M:IfF&D Smg,DdDTX5͑e>ymΡF "7%DP<7[BdBdSxebQbm:FJQ%7rN&ASQ F#Vp%SR(+yR|MpJL nܙ*Tz)9*b H6.lRS1+trSN8bgUXpGG9Q82"Ud+P [;Hzj9$ SnzY yiEe@?>yqPzq{we!gb>J{ÿͷ7 nד7Ksw^O޽z2p)cS J5T##+~|V%P7I& ۽7})#6<`ge zBGO&Ö%gefGHJl;* )!SXWqFLaX y'B\&C6woOxC9gw*Ǵ@(s@eXϜGR+y@p};{G1OA-i2rddA'52*52RT ,l\+-GGFP!pJRP?$@IrsAct(4u-]iQGPjCOqV?Au;;QG$]k.e|L S)[ܒ$}VJ;qDWIO3+ OAg/$NR+i}S;p.vJv#hJI$VUJ?u-")"猏Wrazm;Ǖ,W/sA+RQ(LG. Õ^J{d+( >}1!P80%Z.p%ɧ[GGs܊D%'?,tB "T`hWy>f aW:ɸAYq &FT1<*\I8\aq3l)J@)S ܖIe%iobRkC`)D >`w=7B `p/~28Y Wr`0(䋷xw6ƕ\,5nϛ+wAg^3Wyaf2CN:(Dnɒ3E~v(Ev9#hH u M+O2Nt`¹p&YOuy!s{@N8Q 2 <6qý@#) yp JI2P=ݻ&-CF9euo5Cbb1gH| CݐvÙTZ;ڜ'$kYQynr@!I(~zuQi%J|^a4ME5@Jp+Uʧ@hn]fϬ>wg22;wmmX65 [{vvIiXJ(i{@TK"hy)w88\EG>jxtp)`JVF5f ̬@xȠ,)e1RkG6nL~]>sB&(p:9V׆8k>b{y/0lR_0Bqݺv$"СF,)$Eb&D zpD_<@lyu@jCY/K) Lk},JM睈di(*SJUs LY0)1;t1Z(S9ӋN_b*fX/lHO*PFRbm PQD~N&/pZg56No'㩋$N:PPL0Uxf1VDGq W2p1vjƉ"w2 @3:D!5*ĽWw(ԖG{-sؗOA}/΁-6N|Yw4<1!JpRo-xD3^{&{O%4kW+]it8^S=g9'vzjMV*t?FyZz^GW?bGW"tkH Q;)1`ժ=Ȧȿ;t \K3bP@(F51c( >Lt!FnbN0l 8T*0E36z,f J!{=Vϊ(C3Uro-y4>h5$}6"BG@_毺 Ĩ=Ve^P+-Fi^YZ~ l3Պʱnhn*7\1i>7nZuܯ]i[uhG#17AUZMvT+<^57CYZ*FH$\}J={~o+_A2Bڍ`1ioМS*X9~}SɵC>%}~/05Un bĖr)0"RL@xOx!Y.׺Iֹ=/~o\ %,MqB ѰLPefJX3ߞt y{R'aP'}W/nQr'1QZhw[/7G 4rIˤ*$ :&Q}h$3VڥZM*u]PrxIXݜA0/e՛=ܻ֒KxtuM9[SNPW _t/MR4MZQ;5y6 V@Uis+#$,=aGI8hݴxԔ*f9jjQ'(:=ioRQu +ZT|~I&N0nȇB%墀M/)Ax"64 0T_Rí,Yobc1)S(P) JSDnc ~C &0!X:s`@  !*ʌJ #Yy#dTZ3%NRNhFk[7u=.f"',FF@ZT5!p\D;S0G9H8220c1h£T̲DN2eFpձ^2'InV` |Jc;O2z2l6.]CC/=)=dCO`Ƶ򴸶E6c<?e3VK)OV05Zz: $U%* =X̡D9l5xm mH`a*P1i)}S.iJPaO!dfO ks[;-ynÀUPKlXΙ.D18hྔLSU2׳4_E* RQؖn>S|`& "G7?5ɾ}1xs%}CCЕ:9oKFOi mD@hLL1 v`PrԺqQ<fre@63!J<SC1T߿7?Ez/rAN"viy_ O3_l8rb/s5+}1oV=oƋˏU`@`@ƄUЇSfut,{s8D/R*nL}ZC+tZ,.׻SyN.'f3;]1!\ =c@=ܚ*h-;P.kFϏ?}[t1$zfD? 5ПW8TKŤWOYY+9 k]0iwur}Gڻ{;Ɍx_m2iS$hܨM<}e+{N*%Iy7>U[pHԬY kܱlCَjMn__ʲw_}~xy1;[7^7p &=6jXi<-Urs`3ˇ{q.>k$&uݝQlYx֝^9N&ӸzuuiLWI aQuoHr4"(c6U?Wr t[R{W@o"q/HZ݂^i}0֡ۃvz g 7޵W=f|L||N?N`(1F5Y(5Zr~fEmc8s6P>L`(yt9W-J,ڪŐ/ 懇yw^;3/~(i~!|q:VUi`~;a,ΪWu݈kK 㚑1g@rcibzKELIyu+NAC+F_VCc[%[ h/2#M*~֭' :Nwd^10nun$.dJt)qΘz9}]ϫ88NHrv6mfepųwt`iTlWkκ0ؼ!xJ!5Q~Ԭ?5d˯ktIj޹`S- *& 8U,@sI1$3#Oac& &nʼnU O|P6 vDE_%43{4fqo2Ťx2(eAJ>v V|-$IBO@'G46;ӘE!O:Hk}#IYi*1u<*򋈌"ȗ,lK1غ9f˭V{ V*r{ú-~~?XLr`c-IDɒ|9,|jt5jŭ4F43âD$z=lv~<)<`(]aQqc8"ef d~b4ۂOuglg/ G>/s3itDb`}`eZj1m:s{#Ofo1R::6)K:1e%ߐ,"Fmն;mDwYõǵH?](u76E'G5`>`N6b@ey$2v&USV;"}0;!Wl2R(ň39d:GXqCZ._ ;‹2m<gs$az%')/U(0eWvGW](F~j6vI1i;Wb gS.V\>hz%#NW>}3 q:&ʮmBJ(`([;GEBI2 2zROf+:H.v :脑*&͐9DJ" ͜t)cIGF2*~x][ Մ~֙z_]]4<[rs›N3Zh]uF&q,ʰmeuZ3H˔ȨO~N[%[#䇩'ś& -M(%wE/tcV鮕\2q´q .8IXrN*e.8Kn x)&*14"@4ƛx (120Z+yF)dVs-Or(6*d瀔*t]\k8%vHBR| ;U n{˭X8~#ώPkgU+]\=t][ m[T,Ւ]VB.1QXy@R kULf6sjbFM6̄XQ duKAV,xX]N+fDɵ7zr A g/mXܝW+>MZb /dQq cήF󔈺&F,)w!\LP>)^( 1ѻg`[B밞YZR`~=״kPvԥvW zxvruՀOEYܰD.(}JXMOD wIvNXP;wj[,гhueF!\fK#\k岷YqrhsVP1Z8*G?I'M RlZU*g$&n/Dn|-a"\_Hh/ p&4{$#ЊĶG=AJoyDQ`dX1Ə g!:̪}Mk :dK㜎gX9+ӨK~gHLd>i-9lImz=p[vD)`uJr y=yr{ =FPLL A FR%d:fmE@ۃ‰얌K뀪$8J+'-ݫk^%Xsn%kjㅡ%wɪ`' CKk@Fi|_?9ә(0E~xW fI/Y&;{a]?<PΉ৞ɲn9VGnDCA ѡD*Z֖mĹYKc\lg'z}XL_-;L2x` I$.gCDAЇ\mYTآ>'Y-͕ͱO/ s}sL!>\+e& %Z87 {.HYEl UۛRm-_8&2ڊ2)؛a OQ2{9L/yiI0_FM8 E %+ƙ:A$eg`,$h$Ts5 nr$L^9 BT^/r^/..szug 4O~>;~di@~.٪0\9?xz[q4 {c̓ 1GǏ|vB$r<ǹVN-\ny8\nך:+Gp ؛?Aݟ^ҝ߮QPne'XsÝ,aoH =PK?qrct y>h:+Vi\fc+N3F V+`klgd~lGDra IcCج/ %@vcdHRr6ҬnLc[ĔšWƣJ)GkqWQc|Ѹ{U-$qty5OX(Q2cz8f6}Ǽл4ʓ%r:闖[֤:j3gaq$/1`ӿ?&Ʀ~Do z,yc!D7m} VmS1M=,,r3d-6,䕛Mի(<̳yU1S鯧W<ޕ5q#藉oQ82qtloC'PP뾖dUdwjXmK2L$˿M|(: AuҼrW , ~R`-v~|fgT1yZՠ$RjEeP~7saR,ֵZFJuյ.U^Ktz̆{x}02 f5"x"aF5&VZA.1O.]PI"pF4|jQC͉v̊6LlģzTW{ݽP _)Aw\Xh%gO^O\ѯ뻋3FaHL `J&\*M)* J)Pg|:[T<<1FʉdTȵ6v77YM Ra[X 4@|f?K~ލUQՊ Lb=׵6DًfoXQg EG *ȅڇdy˯fw@ȃC-$n~#.?&좺q4j9^mIz? rUKw,ǔ@MpΑE7#:m$I1[:4U!g[UkD>`)e5ؠ Ji"0X3 >76ic`KZ˅|=649!rکdC60+d29IB5`]C>`P Қ{یfm8b$#m7HI;PZ $O5zp)n&>AR"Qb>Mz4 b!l>2 ]:W >W fˏM塢FQ(xAn# `r_[rߟjIT-u\'5 Pg:T7՜_U01jbE:l/RžuTkʘ8zvjI1tiAJۓhҙpBi2ݙ'3O'WV:%ѯk֍QN Euھu;+Tg&Z.4Ot>u@1XNw4n"8 |n #['W:%cWETqWŴ槺*r8?g~d\Pͤ=Ws6>yGRnY$*eL>Y I;Vxi@eJ qx ɩr  Vp*w\KND u ORid`.5Mʂo$Q5q j֣ZGJX#p~Ua][_ӤnV$Dȝ~x.1l~Ւ]]]ŠuUO+)軠H:ʱTg߽eᵫڤ>5D 03~ZdV*x$HuYQ>MɈGt>[yzr딷lzĆJ x1FD|ߥIn/_rR/߾^ $-x{e[NC1 5%աظf.#1Xw,_WQDԂ݈4u&uը/q!.D8Ct&RN)Z?#H43# "V3"w\f\+k'-ҶzS(} 9.Wdo m@_kQkE S .X:PP^0P.)uܰn(4XÖ%b$tζ)H&z^9N,inW)_8"1>\ClXrnF_7[5oKj$OCDߒےoŶVVf"kJn[\褱.&^Ǎ U eu]lNXukpzlĠRK==mڐvn3R)RatyttfN1]X@J˯jh C&0 ,zG.j]dZȀ3 UANkp;z$a@K'zW#3e6Ip鼍[ p}[5j}&痄ײ}7-֌2z4W}YMup+'C ak`.1/qś>UR}@z"x.5^}m}QxIH1\qU0)X!ђU@8 Zs0aoTp!8.T9Ơa6R{knv#iLWl_ rcp]IN|K6>4|C cQ쿬5no*ȍQK䈚 O n'rdx iAsW yQzϴuy^4x[g֧=XUT<C޷GnRא༨jɵuw3.7(Yk]{SwQZl( s%8}L"޿͙43T^65b+Z+)ez B۫Xz!Pڅ~;,;(+ %;`L{5mAS͐;P4NCF AGQl ,ܦ5i rZowII) a}ãdx!`-z\{HAJ^ ‘Wr]RК*x{8%J*Ry)2M+Q*7rs"NHG]s^@^ʼn XZH@`Z[N*d xta]@zuR՛4:B=&$6$j`l7@l S;JaBқiEB8Ä7JI^b\Z7-s `Q>X`ۦQTGX8B@+pd̒i<84xz3 ȁYp29^\w7Q`D5W:.h C.GeYB)贿hCjg b~үk9e)}CjbM ˕c \JN`&τ*BEf<\k;n=.mKS̴5*v^_nBI) M(/~25I!A/_URMRcf(a3u KO6 L\JoyXJ"D3Q6MdT2I9:U~̌Μ8̶c<./[)Ju}MM[T{dS<`C y,/+Yj&ʕTiJXkWOrM"HAJ?{WFo=Z|+𧝹7ݹ4βi$-J-;ӭnJ. VͧX"UK3Ϥ#Ӆ䆙38?C jOx1E 43m6'>I@H{ }G Ǩmij/G >F_Q9*G t]l) 3g1yIe[uZx5zQe #oC0&ei3k`KLd|hnj7bӞw|̲L!OU7gZqcXiq@Szʿ?P['\Bk.ΊC5RZ6 2I JQT$!H V05_WfQ&ë( u)2@r=l|JPsY#NMl9H`iEޗoHg' -B X2rZAΙ)R\ m,D2!v 鱖ү"Z)W<ox&aRܪ !&ꥫohVٺgY$i>}*g}7㹻#|G1CvE(}4tت㮔Ry0BƯO;FJi0smDfȬ?n+StR͋'ܬ0KFqnm&I}ŇAɠ'VE2⧫ND6 3yקƿ홋EE^]<`nIX|l ɮ" ^)d !С0Z3vgϛ=-Ć3`P˞|H.P{YI9,.JR b5gV3rNg#Mg86/xo(?"ڱ3 Na+n-lvixAKR边3FRm:ވ%"^芇q:<昸FQ0-rdЅFx*'E€g0|ݮ+6i+"]΀@q>:kA18R9 :*\R&$/YeLV3tʥ/ܺTU~be C.J#GnG&$ @JVP6yjCt7@ݠF٢Ofy$*SJUDXbLm](->qs?OLPYlu5(IqIDyR ~/I#ViT&W=L'3I(!glG%Ar͙$bdWF oHqɘUݮ뉛SviqVvڊ;DI'% t7]Fh,J׹Bk6fdW[]-s?0b&Z'ZsI0蠇(>ih2viuzi@E[AdBBf7zaF&(yE:-iyˉU;+ ^X߇v T8z^p;#A:<"#SB6伊*gw^\) }U4RK5; \р-,NçQ]!OI_~sy .ӕ ?^hJ~%Sd;( \WzÐ_ NmJ0fV+BE\ku1E8O\ [}໨BR`ߗ)U0R*.KR)$# Vi醼;;Y='Wui~;}1n)oY41;"@68ن-،LӲФe0JLndw #I% #}ڍV5D.^*gHc/F[rh2fJШޣL w{{MR?ʛQa!g9E=LKbҐ $MJC˝q?N:eHv%R.IGdg? ? e\qK}E'YBROᆅ<H(Ȥ#ɖ"4^V BnsD@yN@*HP9'QPEXHf(Z>-kw0YD/M2}NOwŅأI"ݝF#pMt8u &oUB85 $`1XUϝichrt)(pQĵAk$)Ez6+M8WhKNgUy/a<ϑJE!7WBОڌ87JW]KV[Ũ\,)b 9 ?QQ!Q0Fc`iD6mfe Pz|FO ԺdUчq}3p 6׌5( u*]uHW7PC5&a;$k`^jw|3z|6!hCF *rbD ,*5!p\QBh#sQ*wt,eT{VA%MƠ 2S]%>q8J3U%7ߧ^q{V擒o7G_Jψioa Y;2?i`;́ W/1;/}ɚ6i_ a+gP$?as$eߤ}JVGfW˄*w%WRI='Re;1Oa͞癱lCGlMfr_}9uO׿߃cMf9:P{9sq0U䊽[a{ ]$~~$A°#vE@jl ?ieE]@rfv+3nf.7u>v5SŶ(Ѝ/Ӻƥd.ŧt?WW~F:$NQ3Nn )kD\Q?js97 svϯG-"(Yb`s;((ORU]]TP>z`fg"aDu{gTNژ*´EwG^}(YZ~ע> d|(dmĭ=ħ{[;|ȝ]%J6)R6 )@馻3C"%nϩ:_;DźDW܄?7Qje=Zqϊ|5R{E;5{V>nR}n]qЄuFzTїn[7n}p΢x֩|_}MЇݺ t6#ijh}zC,|Eb]O:f>CrXdz2X=Y8W\`b@B?"cZ&X K1*Sَ- vˀgsd0؇uB aqB\@MR~mTO?~TyU.dG"?jpgX-3gi T֯*I(ؖM>3;:+%Vj}O/#ԁY1'DmAJ/$ JAN兎4'\$231te'JcB1ME;i6XPMhn7OEJ^x=zp$#Ƹ]Ů}ju:_$Wˋ~>,#{?Bzֿ}-ޥ 6SzÁ _~i=|S}Npz1!5Oe ti6_ yJ'LXQ\R2,OhV&(jEKM^-0I 8~9<= g6"# x9r=;EIxDl%?(~ZfMu5z ڼg֛>Rd50Tb{GDtxp/K@ϸhRj( H=G\LқWje$ S2~7k㖚?ֳn˦bYpcKCyoQb !r=O)'nvh|,[C#R|n4O<[I<2KTNrp'O4SYEfYf$-5.m ac33< &Kc96ҎEwԬ2LpLIЁk*AHw"8T^a6Ѩ:F1۫IcWƶ(~78dHjwbQh Rc fY1m`\E0ttZ%U]Ш5ڄxj0(_rwcH诫Zďr<%^ǯ7K# ,#o>~P.8P6_.~k|pgg6@(sRkvlj;ń*Ņeg 4׀.&n,zl%ý)`=d8?<8_#*m_O ,F rpkEoлeQN\H{x!پSYXG~Y2&%! (8l~t-9TKeYϺ~ZH% 8{vl0(F5@7@4J2/6qk hQE׊Gpcˍ=m|[fAgOfXgtĔ[56Ba+n s۟vAb_}/Qn>~PQ˖!c;5o R06zCֵC?_4n zũ}4nEQ.6|cCԗ Q0-^qs6#[#Ę{4$r޾F`HPrf_ڥ{صKK(\}_hJ=rV:ˮem7|dS&yQGEmPs+~|C]KCu˥M;g<{׆q];W]{.]qФ\hC4/尻԰ah'kWP-+j8h:]v{=oikvXP!?8v)Cf XonEؗnj8g{<0QN sGfGϻ:5YakGVkntmEIzK9*5{"|IGљ 'Hr]$:Srtz۟::E9l2*(TLУ$+u>A `")qq'":sE`Oy6Z-O/*W΄Fro#'>L2T=!]A&u8^ `4xr|%*`PD݈\{2^~sd7J#Y%a^)ѐHY_"p.DKɭEͳ OgbFkTvN2x"{ !F|^4]"mhO՗y%e <(eb$9tM%__+K`П~AO7O>.*QF•ReD%.DMQA Ajvv1ZT[T'wآWņn}ъ^ג̌?=bR/s!~O\ҜNCIy7ˆ3> i7v ƒ8 HO+PYT-}!85>,{*E6#řDv`5a`9*q}~y3xq|QnG/( 3 ЧG/FOY >B&dHO(+faĝ7Y(U&U*]_C'B)'YH/rVA3$(OE$#$#(ݴרᴕF(reԱRCu{e:df9YSNsحTre'vH@u7<)Igt<%Y TSJa&P#д^-2(>0A.bkQ $X؍ݏ@ :yqpqh4i1*[41!$0~D_R2j^罴Bq}{;%-xNö-54kO I#ZmPb[˂ϻ*mF6h(5N) zoJ T6jDRR@ ͣhl`[w0vװ{>خy dRc6\uB3]DKllי96:)JM [B^:-ZJw<p5Vƥ$6<r:6ujwn>ڱ6{* <1,QYo=9jSwȐʊ`,cA(JxLF8b7D+YL6x}f( /_;C.Dd!߶VeÓ;Fr^xMͽkZ{uo[XFnt ͑iUA4Jl'g9X&(eQ&Gʤ$ MFHNees3=İ_%r~~JPe|Lj49"Lå EщxmPKʭdBPJrPnɯ2y3,f"H:zd-fY گT_s~>IFDm ++I*HmĤ/RxЃ '>#L:E1G^H>Vd疁{ak@55]#lͷmkwH_r2nƾC ɗBh'`-n˒VAXr?cdU4xnFSފ-Ͽ_lm2Wi 6h`z9r9'!/P0'AX;90m31hEu"_T;1kד,f ;2 b0F,FDqY+bҔ/tPX//` ra 4#/-)5ak]+)#Rժ/b _R76\B97/NYZQNIe0]ZM Kעp.և dᷯBFʂe<Jy2>]˰ O}{,([ɀdh]@`/xK I7oy=Hzjy<( uc.ni:V3JvOcGptw-1?ULPj R D{n" 1'F@JDLyQehd& :G 몈CQ/ \k R ЇԒGz<"" BCpגaLLjzu3:Ri}X{Nѐ.ry}5/2@DNrK10Љ8k珨V㬃ꠔI^ AA.gp .ߡ/FWW[W /g>' 6ӎ aE_}*MH-ӕ[kqAzӮ>4v$as+}$)?t,Kyc%b k !X(s7 (rA4 zQͳb}e| Ϭ{ja']BRig՞4As!6>RŽ0*zD(Y҆Fb򛐱Ѡ.˽uda;:AcC{YRXcpA2 ^ H,#g=eNaˢ̏Ԙ{px/nFhRL<6^$tёHuZ]Nc%wݍ ft-_ jKOcVq#egGmRܖάsfCTꂁepÊN82E3SmE(h!S5-T(} bн6 T VcD*X$eU9yB)`Qe1>R Reͱ6ATSrV.Ri:`K|jzQLJY p 2H5%8Jוu$JR-(=H&}!ќV:0|XNm؅ pz ( G8*1*wGHssӊQQkYpk:;&s> LY,[ߨaz8ޏ~nKc`FCդ*_CeˤѰcA6b4<ӻr2et9s6+R*f4cЋHi3~'u=7)VRؿ&N/H؆Ne7wO&~ѓ}=HwtWJ;. #7!x(~6?^{7Wj.$8vw2>/Y!OߋrWnQh1V ?ES}!|_zo.G@3_@8gkݎFK+h؄;ar+{w2μ`dg^lܭ@RMYhDu,HVM5FJFW1sL5(xc|Gi>Mza<-Vzx8ZBbo|/q4 h}~M'" gn4,4OϡUn*_ONu~8ݭ_4}ǜi|x1+4:igTӇ w j3ݔh%3"=rmƌ@O+B^9DaJGMnMec9 gݚ&Z6!))էTJٶ`η9eӻKG̮{D/. FP54oIA$i@}:p ϗZYh:Jޮ4=e4 <ǕJMCiI5BRP<ߙv;2=lJC)u>_$9C;l&cLoiK<6p@8:ec BQDlncŹ`JBatjvJfx]yZT/S!-vJEJ();MT/S޴:QIJy0 rҒjI/8C9ˎ5Z2ZLFAҮ8$}u9zqB`5moօFF(TD S*aKtYu}Ib[UzYރ ܣw&<ȗ)x 4û@'] @=8ٓ ѽyTn ,nR#*lpNʈcO7>fk̏4q U ޻>i/L.>|<ڔ?V@M ; X`XJ͂@Gl,U)l Y[iYXJ}QfJdJi T[-:͔R]3|~gJT%q_GD}6pHh*.`_rTNKzhžoA%}sKvig-k|o]?}8>!ٺU{XCLeb^be/~|g<[E6M9#eMfF#.dK-pU{7ZVLϊCIel Y'_&_s֒(r9Z[釭өx*S9Gd;`mNwu Si&Zy*czu6*%H]Jl;R?D >PfmG`KmPPi^ڕ@;l6$ }窕$4-DzᜠK Z퍌L.B TAV8dԢB<[[+x`(%?yK5w,R0L\J)梘/RɇG݁υ/n-|gӳݏߕ">,DH$Su/@9cՖ$r,u61Z3A4͵ᑪe e_SZ6f|ɭ%Mn҉} !E{ȡ\hA܆ d$ѲRTs}iNo6x&|J͗Wnzpc\ vV8G.yTR(9o%׀J9;?󉌴3s5Ox7~'-4&3²jH7z~~wd<.aV]uO91q{WoTNR.,Ն>3Js[=zpq-l8gEV=(\~"xLHFIk>u_'?Vjׂ !!^Sыߝ=Y@dd6oCZ !6 Ph&{Z<Ϣh9~1?(M(-fs~tGgcW`n2$xi80nzW_Aǽ/=:~x-KD}.`iѺd8ju7l$eIg%_hr쵗1Xb\DS.qmg+7Wi%r%83 t6EJFzۅq#m5o#3ibN.q W{IX[rEijQi:u_Jڒa|raϮdRFط^YM[ᵅSewB01 S=8_&-Cd?&&?yVi?XL1k<\kTOx@w'@YiUvl{utt1زUF52j bfŝg[Vj|Zjk)nTM#X޾ Rj7}GN~m CȺf:Ѽ Ru.)g٬[󳀉֭ r6t|l$uk*4}G(J][%֭ r.aJ%b*&Dik͛Z<5dvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005156450115145557001017707 0ustar rootrootFeb 19 08:43:37 crc systemd[1]: Starting Kubernetes Kubelet... Feb 19 08:43:37 crc restorecon[4674]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:37 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 19 08:43:38 crc restorecon[4674]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 19 08:43:38 crc kubenswrapper[4675]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 19 08:43:38 crc kubenswrapper[4675]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 19 08:43:38 crc kubenswrapper[4675]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 19 08:43:38 crc kubenswrapper[4675]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 19 08:43:38 crc kubenswrapper[4675]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 19 08:43:38 crc kubenswrapper[4675]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.822107 4675 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830564 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830589 4675 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830593 4675 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830600 4675 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830604 4675 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830608 4675 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830612 4675 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830617 4675 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830621 4675 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830625 4675 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830631 4675 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830664 4675 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830671 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830676 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830682 4675 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830687 4675 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830693 4675 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830697 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830701 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830705 4675 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830708 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830712 4675 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830715 4675 feature_gate.go:330] unrecognized feature gate: Example Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830719 4675 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830722 4675 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830726 4675 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830730 4675 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830734 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830737 4675 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830741 4675 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830745 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830749 4675 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830753 4675 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830758 4675 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830763 4675 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830768 4675 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830772 4675 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830776 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830780 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830784 4675 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830788 4675 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830792 4675 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830796 4675 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830800 4675 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830807 4675 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830813 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830817 4675 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830822 4675 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830826 4675 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830830 4675 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830834 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830838 4675 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830842 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830846 4675 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830851 4675 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830855 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830859 4675 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830862 4675 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830866 4675 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830870 4675 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830873 4675 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830877 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830880 4675 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830883 4675 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830893 4675 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830897 4675 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830900 4675 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830904 4675 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830907 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830911 4675 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.830914 4675 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832122 4675 flags.go:64] FLAG: --address="0.0.0.0" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832141 4675 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832157 4675 flags.go:64] FLAG: --anonymous-auth="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832171 4675 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832181 4675 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832185 4675 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832193 4675 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832200 4675 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832206 4675 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832210 4675 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832215 4675 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832220 4675 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832231 4675 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832236 4675 flags.go:64] FLAG: --cgroup-root="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832240 4675 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832273 4675 flags.go:64] FLAG: --client-ca-file="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832278 4675 flags.go:64] FLAG: --cloud-config="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832282 4675 flags.go:64] FLAG: --cloud-provider="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832286 4675 flags.go:64] FLAG: --cluster-dns="[]" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832298 4675 flags.go:64] FLAG: --cluster-domain="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832303 4675 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832307 4675 flags.go:64] FLAG: --config-dir="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832313 4675 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832318 4675 flags.go:64] FLAG: --container-log-max-files="5" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832325 4675 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832330 4675 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832335 4675 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832339 4675 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832344 4675 flags.go:64] FLAG: --contention-profiling="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832348 4675 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832353 4675 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832357 4675 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832362 4675 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832367 4675 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832371 4675 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832376 4675 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832381 4675 flags.go:64] FLAG: --enable-load-reader="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832385 4675 flags.go:64] FLAG: --enable-server="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832390 4675 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832399 4675 flags.go:64] FLAG: --event-burst="100" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832406 4675 flags.go:64] FLAG: --event-qps="50" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832410 4675 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832414 4675 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832419 4675 flags.go:64] FLAG: --eviction-hard="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832425 4675 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832429 4675 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832434 4675 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832438 4675 flags.go:64] FLAG: --eviction-soft="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832449 4675 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832453 4675 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832458 4675 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832462 4675 flags.go:64] FLAG: --experimental-mounter-path="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832466 4675 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832471 4675 flags.go:64] FLAG: --fail-swap-on="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832475 4675 flags.go:64] FLAG: --feature-gates="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832481 4675 flags.go:64] FLAG: --file-check-frequency="20s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832485 4675 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832489 4675 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832494 4675 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832499 4675 flags.go:64] FLAG: --healthz-port="10248" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832503 4675 flags.go:64] FLAG: --help="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832507 4675 flags.go:64] FLAG: --hostname-override="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832511 4675 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832516 4675 flags.go:64] FLAG: --http-check-frequency="20s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832520 4675 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832525 4675 flags.go:64] FLAG: --image-credential-provider-config="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832529 4675 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832533 4675 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832538 4675 flags.go:64] FLAG: --image-service-endpoint="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832542 4675 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832547 4675 flags.go:64] FLAG: --kube-api-burst="100" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832553 4675 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832566 4675 flags.go:64] FLAG: --kube-api-qps="50" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832571 4675 flags.go:64] FLAG: --kube-reserved="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832576 4675 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832580 4675 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832585 4675 flags.go:64] FLAG: --kubelet-cgroups="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832589 4675 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832593 4675 flags.go:64] FLAG: --lock-file="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832598 4675 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832602 4675 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832608 4675 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832617 4675 flags.go:64] FLAG: --log-json-split-stream="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832622 4675 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832651 4675 flags.go:64] FLAG: --log-text-split-stream="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832658 4675 flags.go:64] FLAG: --logging-format="text" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832662 4675 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832667 4675 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832671 4675 flags.go:64] FLAG: --manifest-url="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832676 4675 flags.go:64] FLAG: --manifest-url-header="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832682 4675 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832687 4675 flags.go:64] FLAG: --max-open-files="1000000" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832693 4675 flags.go:64] FLAG: --max-pods="110" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832697 4675 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832725 4675 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832730 4675 flags.go:64] FLAG: --memory-manager-policy="None" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832734 4675 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832738 4675 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832742 4675 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832747 4675 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832760 4675 flags.go:64] FLAG: --node-status-max-images="50" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832765 4675 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832770 4675 flags.go:64] FLAG: --oom-score-adj="-999" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832776 4675 flags.go:64] FLAG: --pod-cidr="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832785 4675 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832796 4675 flags.go:64] FLAG: --pod-manifest-path="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832801 4675 flags.go:64] FLAG: --pod-max-pids="-1" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832807 4675 flags.go:64] FLAG: --pods-per-core="0" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832812 4675 flags.go:64] FLAG: --port="10250" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832816 4675 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832821 4675 flags.go:64] FLAG: --provider-id="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832827 4675 flags.go:64] FLAG: --qos-reserved="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832833 4675 flags.go:64] FLAG: --read-only-port="10255" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832839 4675 flags.go:64] FLAG: --register-node="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832845 4675 flags.go:64] FLAG: --register-schedulable="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832850 4675 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832859 4675 flags.go:64] FLAG: --registry-burst="10" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832863 4675 flags.go:64] FLAG: --registry-qps="5" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832867 4675 flags.go:64] FLAG: --reserved-cpus="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832872 4675 flags.go:64] FLAG: --reserved-memory="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832886 4675 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832891 4675 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832895 4675 flags.go:64] FLAG: --rotate-certificates="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832899 4675 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832904 4675 flags.go:64] FLAG: --runonce="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832909 4675 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832913 4675 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832918 4675 flags.go:64] FLAG: --seccomp-default="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832922 4675 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832926 4675 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832930 4675 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832935 4675 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832939 4675 flags.go:64] FLAG: --storage-driver-password="root" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832945 4675 flags.go:64] FLAG: --storage-driver-secure="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832950 4675 flags.go:64] FLAG: --storage-driver-table="stats" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832954 4675 flags.go:64] FLAG: --storage-driver-user="root" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832961 4675 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832966 4675 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832971 4675 flags.go:64] FLAG: --system-cgroups="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832976 4675 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832984 4675 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832989 4675 flags.go:64] FLAG: --tls-cert-file="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.832993 4675 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833002 4675 flags.go:64] FLAG: --tls-min-version="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833006 4675 flags.go:64] FLAG: --tls-private-key-file="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833011 4675 flags.go:64] FLAG: --topology-manager-policy="none" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833015 4675 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833019 4675 flags.go:64] FLAG: --topology-manager-scope="container" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833024 4675 flags.go:64] FLAG: --v="2" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833032 4675 flags.go:64] FLAG: --version="false" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833039 4675 flags.go:64] FLAG: --vmodule="" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833046 4675 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833050 4675 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833185 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833191 4675 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833197 4675 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833207 4675 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833211 4675 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833214 4675 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833218 4675 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833221 4675 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833225 4675 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833229 4675 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833233 4675 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833237 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833246 4675 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833252 4675 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833258 4675 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833265 4675 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833270 4675 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833275 4675 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833279 4675 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833284 4675 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833288 4675 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833293 4675 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833298 4675 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833302 4675 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833307 4675 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833312 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833316 4675 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833319 4675 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833323 4675 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833328 4675 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833332 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833336 4675 feature_gate.go:330] unrecognized feature gate: Example Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833339 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833343 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833347 4675 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833350 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833354 4675 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833357 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833361 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833371 4675 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833375 4675 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833378 4675 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833382 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833386 4675 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833393 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833398 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833402 4675 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833410 4675 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833414 4675 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833418 4675 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833423 4675 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833427 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833431 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833435 4675 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833440 4675 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833444 4675 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833448 4675 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833452 4675 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833457 4675 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833462 4675 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833466 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833470 4675 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833475 4675 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833480 4675 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833484 4675 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833488 4675 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833492 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833497 4675 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833500 4675 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833504 4675 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.833512 4675 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.833527 4675 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.850182 4675 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.850237 4675 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850468 4675 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850498 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850510 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850523 4675 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850535 4675 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850545 4675 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850559 4675 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850576 4675 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850590 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850602 4675 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850612 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850622 4675 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850669 4675 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850682 4675 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850693 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850703 4675 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850713 4675 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850725 4675 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850735 4675 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850774 4675 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850788 4675 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850801 4675 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850813 4675 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850824 4675 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850835 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850846 4675 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850857 4675 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850867 4675 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850877 4675 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850888 4675 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850901 4675 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850912 4675 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850922 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850932 4675 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850942 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850953 4675 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850963 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850974 4675 feature_gate.go:330] unrecognized feature gate: Example Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850984 4675 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.850994 4675 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851004 4675 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851018 4675 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851036 4675 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851048 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851060 4675 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851072 4675 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851083 4675 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851094 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851107 4675 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851118 4675 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851129 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851139 4675 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851153 4675 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851167 4675 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851179 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851210 4675 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851221 4675 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851233 4675 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851243 4675 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851253 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851263 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851274 4675 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851284 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851295 4675 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851305 4675 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851316 4675 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851326 4675 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851336 4675 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851346 4675 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851356 4675 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851366 4675 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.851385 4675 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851766 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851791 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851804 4675 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851815 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851827 4675 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851839 4675 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851849 4675 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851864 4675 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851875 4675 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851888 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851898 4675 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851908 4675 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851918 4675 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851929 4675 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851939 4675 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851949 4675 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851960 4675 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851970 4675 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851980 4675 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.851992 4675 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852006 4675 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852017 4675 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852028 4675 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852039 4675 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852050 4675 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852060 4675 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852070 4675 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852080 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852089 4675 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852100 4675 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852115 4675 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852129 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852142 4675 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852154 4675 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852167 4675 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852181 4675 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852192 4675 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852203 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852212 4675 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852226 4675 feature_gate.go:330] unrecognized feature gate: Example Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852239 4675 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852292 4675 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852305 4675 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852319 4675 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852331 4675 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852344 4675 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852358 4675 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852369 4675 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852380 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852390 4675 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852400 4675 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852411 4675 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852420 4675 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852431 4675 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852441 4675 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852452 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852462 4675 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852472 4675 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852482 4675 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852492 4675 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852502 4675 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852512 4675 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852522 4675 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852532 4675 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852542 4675 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852552 4675 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852562 4675 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852572 4675 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852582 4675 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852592 4675 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 19 08:43:38 crc kubenswrapper[4675]: W0219 08:43:38.852602 4675 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.852621 4675 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.853996 4675 server.go:940] "Client rotation is on, will bootstrap in background" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.861586 4675 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.862776 4675 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.864775 4675 server.go:997] "Starting client certificate rotation" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.864829 4675 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.865087 4675 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-21 03:28:13.827448671 +0000 UTC Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.865237 4675 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.890025 4675 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.892547 4675 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 19 08:43:38 crc kubenswrapper[4675]: E0219 08:43:38.893176 4675 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.920149 4675 log.go:25] "Validated CRI v1 runtime API" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.963774 4675 log.go:25] "Validated CRI v1 image API" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.966564 4675 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.972044 4675 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-19-08-39-32-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 19 08:43:38 crc kubenswrapper[4675]: I0219 08:43:38.972103 4675 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.012106 4675 manager.go:217] Machine: {Timestamp:2026-02-19 08:43:39.007994197 +0000 UTC m=+0.635084555 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8ed5780f-8c9e-4eed-a721-460b032942bc BootID:42540fd8-b49d-428b-9ca1-9222fab1c7ed Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:5a:c7:91 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:5a:c7:91 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c5:b0:27 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:18:a5:cc Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d8:a6:d5 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:93:fa:8d Speed:-1 Mtu:1496} {Name:eth10 MacAddress:c2:21:5f:e2:17:2f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:16:37:30:42:7b:5d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.012614 4675 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.013034 4675 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.013562 4675 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.013961 4675 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.014027 4675 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.014457 4675 topology_manager.go:138] "Creating topology manager with none policy" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.014487 4675 container_manager_linux.go:303] "Creating device plugin manager" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.015159 4675 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.015214 4675 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.015997 4675 state_mem.go:36] "Initialized new in-memory state store" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.016156 4675 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.020080 4675 kubelet.go:418] "Attempting to sync node with API server" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.020109 4675 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.020137 4675 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.020153 4675 kubelet.go:324] "Adding apiserver pod source" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.020169 4675 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.024663 4675 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.025916 4675 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 19 08:43:39 crc kubenswrapper[4675]: W0219 08:43:39.026580 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.026682 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:39 crc kubenswrapper[4675]: W0219 08:43:39.026676 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.026782 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.029143 4675 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030778 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030813 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030825 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030836 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030854 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030865 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030876 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030891 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030905 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030916 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030929 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.030937 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.031674 4675 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.032332 4675 server.go:1280] "Started kubelet" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.032937 4675 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.033188 4675 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.032961 4675 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.034713 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.034751 4675 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.034771 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 23:33:18.808636022 +0000 UTC Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.034797 4675 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.035007 4675 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.035026 4675 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.035068 4675 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.035105 4675 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 19 08:43:39 crc systemd[1]: Started Kubernetes Kubelet. Feb 19 08:43:39 crc kubenswrapper[4675]: W0219 08:43:39.035495 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.035556 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.036475 4675 server.go:460] "Adding debug handlers to kubelet server" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.036763 4675 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.036797 4675 factory.go:55] Registering systemd factory Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.041492 4675 factory.go:221] Registration of the systemd container factory successfully Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.043993 4675 factory.go:153] Registering CRI-O factory Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.044055 4675 factory.go:221] Registration of the crio container factory successfully Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.044116 4675 factory.go:103] Registering Raw factory Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.044143 4675 manager.go:1196] Started watching for new ooms in manager Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.036782 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="200ms" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.045554 4675 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189599550e5b2f9f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-19 08:43:39.032285087 +0000 UTC m=+0.659375375,LastTimestamp:2026-02-19 08:43:39.032285087 +0000 UTC m=+0.659375375,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.050620 4675 manager.go:319] Starting recovery of all containers Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.052822 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.052929 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.052947 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.052961 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.052976 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.052993 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053014 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053030 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053052 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053071 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053085 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053098 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053113 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053130 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053172 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.053183 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057409 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057457 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057472 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057484 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057499 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057511 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057523 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057535 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057547 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057558 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057572 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057587 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057603 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057613 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057624 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057633 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057679 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057694 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057707 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057719 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057730 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057781 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057793 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057823 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057837 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057850 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057861 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057873 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057885 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057896 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057908 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057919 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057932 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057945 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057957 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057969 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.057990 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058008 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058023 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058038 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058051 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058065 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058078 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058093 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058110 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058123 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058138 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058154 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058169 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058184 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058196 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058210 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058224 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058234 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058245 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058256 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058267 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058279 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058290 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058302 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.058313 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060179 4675 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060212 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060227 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060240 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060283 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060294 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060304 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060314 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060324 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060333 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060347 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060359 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060371 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060383 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060394 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060407 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060421 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060433 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060443 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060455 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060469 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060480 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060493 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060507 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060519 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060532 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060545 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060555 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060605 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060620 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060633 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060699 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060713 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060726 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060738 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060753 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060767 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060781 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060795 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060825 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060836 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060846 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060858 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060868 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060879 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060893 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060904 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060929 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060940 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060954 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060966 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060977 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.060988 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061000 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061011 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061022 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061035 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061046 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061058 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061069 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061083 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061095 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061108 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061121 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061133 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061151 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061163 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061174 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061185 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061197 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061209 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061221 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061231 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061243 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061254 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061266 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061278 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061289 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061301 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061315 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061327 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061338 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061352 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061361 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061375 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061391 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061403 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061416 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061429 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061442 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061476 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061491 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061503 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061513 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061523 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061535 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061545 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061558 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061573 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061585 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061596 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061608 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061618 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061628 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061665 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061676 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061686 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061696 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061707 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061723 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061735 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061746 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061756 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061768 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061778 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061790 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061801 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061814 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061824 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061836 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061849 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061861 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061872 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061887 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061900 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061912 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061922 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061937 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061948 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061961 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061974 4675 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061984 4675 reconstruct.go:97] "Volume reconstruction finished" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.061993 4675 reconciler.go:26] "Reconciler: start to sync state" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.067923 4675 manager.go:324] Recovery completed Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.079117 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.081451 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.081500 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.081513 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.082410 4675 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.082431 4675 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.082456 4675 state_mem.go:36] "Initialized new in-memory state store" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.098948 4675 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.100615 4675 policy_none.go:49] "None policy: Start" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.101712 4675 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.101745 4675 state_mem.go:35] "Initializing new in-memory state store" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.101917 4675 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.101972 4675 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.102011 4675 kubelet.go:2335] "Starting kubelet main sync loop" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.102080 4675 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 19 08:43:39 crc kubenswrapper[4675]: W0219 08:43:39.104394 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.104476 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.135503 4675 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.171157 4675 manager.go:334] "Starting Device Plugin manager" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.171229 4675 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.171250 4675 server.go:79] "Starting device plugin registration server" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.173826 4675 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.173857 4675 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.174085 4675 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.174260 4675 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.174276 4675 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.181994 4675 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.203224 4675 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.203383 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.205330 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.205395 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.205410 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.205710 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.205865 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.205926 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207213 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207246 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207260 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207285 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207325 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207383 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207520 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.207567 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208315 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208356 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208372 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208552 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208657 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208697 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208713 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208788 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.208837 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.209260 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.209289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.209301 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.209473 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.209597 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.209652 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210440 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210446 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210499 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210514 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210471 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210794 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.210829 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.211823 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.211852 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.211870 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.212344 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.212375 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.212391 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.245780 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="400ms" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.264905 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265091 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265225 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265310 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265492 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265574 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265609 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265660 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265748 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265817 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265869 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.265982 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.266063 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.266103 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.266143 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.274039 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.275354 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.275405 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.275427 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.275470 4675 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.276039 4675 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367552 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367612 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367662 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367689 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367712 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367732 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367744 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367782 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367751 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367828 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367836 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367854 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367861 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367857 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367910 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367914 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367885 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367931 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367891 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.367874 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368017 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368049 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368075 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368100 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368123 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368128 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368157 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368160 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368200 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.368247 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.476491 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.478307 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.478375 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.478398 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.478443 4675 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.479154 4675 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.535839 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.567960 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: W0219 08:43:39.579077 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-83d10838a3b711b6fb1ded58d0d40ec6506bf03964e62eb2aecb55c293b31fe9 WatchSource:0}: Error finding container 83d10838a3b711b6fb1ded58d0d40ec6506bf03964e62eb2aecb55c293b31fe9: Status 404 returned error can't find the container with id 83d10838a3b711b6fb1ded58d0d40ec6506bf03964e62eb2aecb55c293b31fe9 Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.589143 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.612606 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.627908 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:39 crc kubenswrapper[4675]: W0219 08:43:39.630038 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-592151935f8657f1d3f637637756d755ed256df2733b79807babc183befbd0e8 WatchSource:0}: Error finding container 592151935f8657f1d3f637637756d755ed256df2733b79807babc183befbd0e8: Status 404 returned error can't find the container with id 592151935f8657f1d3f637637756d755ed256df2733b79807babc183befbd0e8 Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.646608 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="800ms" Feb 19 08:43:39 crc kubenswrapper[4675]: W0219 08:43:39.651413 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-ba9bf6b986a1d09e684dd6d8f66e1f1948cba8d58e2b8d0375767d96f152d895 WatchSource:0}: Error finding container ba9bf6b986a1d09e684dd6d8f66e1f1948cba8d58e2b8d0375767d96f152d895: Status 404 returned error can't find the container with id ba9bf6b986a1d09e684dd6d8f66e1f1948cba8d58e2b8d0375767d96f152d895 Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.880014 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.882189 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.882240 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.882253 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:39 crc kubenswrapper[4675]: I0219 08:43:39.882285 4675 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 19 08:43:39 crc kubenswrapper[4675]: E0219 08:43:39.882628 4675 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.034962 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 03:40:32.852116107 +0000 UTC Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.035005 4675 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.109784 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"83d10838a3b711b6fb1ded58d0d40ec6506bf03964e62eb2aecb55c293b31fe9"} Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.111438 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ba9bf6b986a1d09e684dd6d8f66e1f1948cba8d58e2b8d0375767d96f152d895"} Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.113671 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"592151935f8657f1d3f637637756d755ed256df2733b79807babc183befbd0e8"} Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.115064 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"69545726747b05bfc50e8f0fbcb5b1874631bb5251685de5923ebaba650cbbab"} Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.116421 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c96f0da121b519e593e109c79967647a2c6cab86bd448d40ab03f27455e5929"} Feb 19 08:43:40 crc kubenswrapper[4675]: W0219 08:43:40.145613 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:40 crc kubenswrapper[4675]: E0219 08:43:40.145717 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:40 crc kubenswrapper[4675]: W0219 08:43:40.191430 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:40 crc kubenswrapper[4675]: E0219 08:43:40.191528 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:40 crc kubenswrapper[4675]: W0219 08:43:40.201836 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:40 crc kubenswrapper[4675]: E0219 08:43:40.201968 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:40 crc kubenswrapper[4675]: W0219 08:43:40.316450 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:40 crc kubenswrapper[4675]: E0219 08:43:40.316540 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:40 crc kubenswrapper[4675]: E0219 08:43:40.447667 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="1.6s" Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.683041 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.685182 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.685263 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.685278 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.685324 4675 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 19 08:43:40 crc kubenswrapper[4675]: E0219 08:43:40.686058 4675 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 19 08:43:40 crc kubenswrapper[4675]: I0219 08:43:40.900652 4675 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 19 08:43:40 crc kubenswrapper[4675]: E0219 08:43:40.902060 4675 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.035055 4675 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.035142 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 00:40:19.130001199 +0000 UTC Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.121151 4675 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6" exitCode=0 Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.121247 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.121342 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.122859 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.122915 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.122930 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.124340 4675 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a" exitCode=0 Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.124439 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.124769 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.127945 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.128019 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.128041 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.129920 4675 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb" exitCode=0 Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.130272 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.137866 4675 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f143cde834aac64e4b7c9d0a3830884e5d494a735ae31f07c4481845693419b0" exitCode=0 Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.138277 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.138936 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.147915 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.148035 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f143cde834aac64e4b7c9d0a3830884e5d494a735ae31f07c4481845693419b0"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.148305 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.148363 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.148376 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.148707 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.148749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.148764 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.149389 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.149420 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.149431 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.152008 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.152065 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.152083 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.152096 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d"} Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.152209 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.158998 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.159079 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:41 crc kubenswrapper[4675]: I0219 08:43:41.159094 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.035199 4675 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.035396 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 00:35:59.519563984 +0000 UTC Feb 19 08:43:42 crc kubenswrapper[4675]: E0219 08:43:42.049201 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="3.2s" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.158308 4675 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007" exitCode=0 Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.158445 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.158528 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.159671 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.159706 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.159718 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.160570 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.160560 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c65048a1f73a3695505b901ea5ad383c9540482d9c10923472f61fda98f8eae7"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.168441 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.168480 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.168490 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.170888 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.170927 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.170967 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.171074 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.171986 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.172010 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.172021 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.176271 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.176972 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.177024 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.177041 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.177054 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27"} Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.177531 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.177577 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.177589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:42 crc kubenswrapper[4675]: W0219 08:43:42.251675 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:42 crc kubenswrapper[4675]: E0219 08:43:42.251813 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.287190 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.288680 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.288723 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.288732 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:42 crc kubenswrapper[4675]: I0219 08:43:42.288763 4675 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 19 08:43:42 crc kubenswrapper[4675]: E0219 08:43:42.289189 4675 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 19 08:43:42 crc kubenswrapper[4675]: W0219 08:43:42.607344 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:42 crc kubenswrapper[4675]: E0219 08:43:42.607454 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:42 crc kubenswrapper[4675]: W0219 08:43:42.663303 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 19 08:43:42 crc kubenswrapper[4675]: E0219 08:43:42.663411 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.035709 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 04:04:25.308362041 +0000 UTC Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.181856 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0"} Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.182054 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.182917 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.182954 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.182966 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.184715 4675 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6" exitCode=0 Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.184752 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6"} Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.184825 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.184829 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.184909 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.184953 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.185589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.185625 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.185654 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.185863 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.185917 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.185937 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.186726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.186749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.186761 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:43 crc kubenswrapper[4675]: I0219 08:43:43.837739 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.036385 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 16:51:33.534578574 +0000 UTC Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192705 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4"} Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192767 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e"} Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192787 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36"} Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192804 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192820 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6"} Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192786 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192905 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.192842 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05"} Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.193895 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.193936 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.193950 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.194109 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.194142 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.194154 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:44 crc kubenswrapper[4675]: I0219 08:43:44.791081 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.036798 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 10:58:01.153869562 +0000 UTC Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.139749 4675 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.200754 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.200805 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.201407 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.202369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.202416 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.202432 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.203226 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.203259 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.203308 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.490080 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.491308 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.491358 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.491370 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.491397 4675 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 19 08:43:45 crc kubenswrapper[4675]: I0219 08:43:45.615072 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.037804 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 08:14:15.13978862 +0000 UTC Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.080096 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.203948 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.204021 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.205038 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.205082 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.205098 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.205159 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.205235 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.205247 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.726082 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.726389 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.728081 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.728160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:46 crc kubenswrapper[4675]: I0219 08:43:46.728185 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.038322 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 20:20:09.827065202 +0000 UTC Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.177377 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.206120 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.206192 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.207762 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.207814 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.207830 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.207840 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.207889 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:47 crc kubenswrapper[4675]: I0219 08:43:47.207907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:48 crc kubenswrapper[4675]: I0219 08:43:48.039243 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 22:10:10.193635675 +0000 UTC Feb 19 08:43:48 crc kubenswrapper[4675]: I0219 08:43:48.233038 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:48 crc kubenswrapper[4675]: I0219 08:43:48.233224 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:48 crc kubenswrapper[4675]: I0219 08:43:48.234609 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:48 crc kubenswrapper[4675]: I0219 08:43:48.234726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:48 crc kubenswrapper[4675]: I0219 08:43:48.234744 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:48 crc kubenswrapper[4675]: I0219 08:43:48.238628 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.040226 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 20:46:33.272169034 +0000 UTC Feb 19 08:43:49 crc kubenswrapper[4675]: E0219 08:43:49.182090 4675 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.211581 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.213007 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.213044 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.213053 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.318251 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.318452 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.319748 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.319774 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:49 crc kubenswrapper[4675]: I0219 08:43:49.319782 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:50 crc kubenswrapper[4675]: I0219 08:43:50.041125 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 12:17:30.490266173 +0000 UTC Feb 19 08:43:50 crc kubenswrapper[4675]: I0219 08:43:50.111236 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:50 crc kubenswrapper[4675]: I0219 08:43:50.214435 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:50 crc kubenswrapper[4675]: I0219 08:43:50.219055 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:50 crc kubenswrapper[4675]: I0219 08:43:50.219091 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:50 crc kubenswrapper[4675]: I0219 08:43:50.219103 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:50 crc kubenswrapper[4675]: I0219 08:43:50.221188 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:43:51 crc kubenswrapper[4675]: I0219 08:43:51.041314 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 05:32:02.062612888 +0000 UTC Feb 19 08:43:51 crc kubenswrapper[4675]: I0219 08:43:51.218584 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:51 crc kubenswrapper[4675]: I0219 08:43:51.219710 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:51 crc kubenswrapper[4675]: I0219 08:43:51.219747 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:51 crc kubenswrapper[4675]: I0219 08:43:51.219763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:52 crc kubenswrapper[4675]: I0219 08:43:52.043330 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 06:15:34.074080549 +0000 UTC Feb 19 08:43:52 crc kubenswrapper[4675]: I0219 08:43:52.134882 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 19 08:43:52 crc kubenswrapper[4675]: I0219 08:43:52.135507 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:52 crc kubenswrapper[4675]: I0219 08:43:52.137120 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:52 crc kubenswrapper[4675]: I0219 08:43:52.137226 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:52 crc kubenswrapper[4675]: I0219 08:43:52.137289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.035739 4675 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.044257 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 16:33:35.765213289 +0000 UTC Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.111650 4675 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.112127 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 19 08:43:53 crc kubenswrapper[4675]: W0219 08:43:53.127045 4675 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.127158 4675 trace.go:236] Trace[1474457397]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (19-Feb-2026 08:43:43.125) (total time: 10001ms): Feb 19 08:43:53 crc kubenswrapper[4675]: Trace[1474457397]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:43:53.127) Feb 19 08:43:53 crc kubenswrapper[4675]: Trace[1474457397]: [10.00176263s] [10.00176263s] END Feb 19 08:43:53 crc kubenswrapper[4675]: E0219 08:43:53.127187 4675 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.176616 4675 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.176944 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.181765 4675 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 19 08:43:53 crc kubenswrapper[4675]: I0219 08:43:53.181860 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 19 08:43:54 crc kubenswrapper[4675]: I0219 08:43:54.044966 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 22:02:50.041125759 +0000 UTC Feb 19 08:43:55 crc kubenswrapper[4675]: I0219 08:43:55.045361 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 23:08:24.617694053 +0000 UTC Feb 19 08:43:55 crc kubenswrapper[4675]: I0219 08:43:55.622633 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:55 crc kubenswrapper[4675]: I0219 08:43:55.623054 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:55 crc kubenswrapper[4675]: I0219 08:43:55.625403 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:55 crc kubenswrapper[4675]: I0219 08:43:55.625470 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:55 crc kubenswrapper[4675]: I0219 08:43:55.625495 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:55 crc kubenswrapper[4675]: I0219 08:43:55.634208 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:43:56 crc kubenswrapper[4675]: I0219 08:43:56.046448 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 02:46:03.504006912 +0000 UTC Feb 19 08:43:56 crc kubenswrapper[4675]: I0219 08:43:56.231835 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:43:56 crc kubenswrapper[4675]: I0219 08:43:56.233053 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:43:56 crc kubenswrapper[4675]: I0219 08:43:56.233116 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:43:56 crc kubenswrapper[4675]: I0219 08:43:56.233136 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:43:57 crc kubenswrapper[4675]: I0219 08:43:57.048392 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 10:38:54.431879859 +0000 UTC Feb 19 08:43:57 crc kubenswrapper[4675]: I0219 08:43:57.158289 4675 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.031219 4675 apiserver.go:52] "Watching apiserver" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.036926 4675 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.037168 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.037474 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.037550 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.037604 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.037868 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.037896 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.037878 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.037946 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.038029 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.038287 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.039254 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.039725 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.039742 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.039744 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.039825 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.039918 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.040779 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.041844 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.041957 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.048495 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 11:40:32.034022883 +0000 UTC Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.068308 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.084248 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.098974 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.107794 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.117168 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.128841 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.136876 4675 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.152027 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.163858 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.176145 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.178432 4675 trace.go:236] Trace[954496153]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (19-Feb-2026 08:43:46.828) (total time: 11350ms): Feb 19 08:43:58 crc kubenswrapper[4675]: Trace[954496153]: ---"Objects listed" error: 11350ms (08:43:58.178) Feb 19 08:43:58 crc kubenswrapper[4675]: Trace[954496153]: [11.350335205s] [11.350335205s] END Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.178480 4675 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.179125 4675 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.179151 4675 trace.go:236] Trace[1973479468]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (19-Feb-2026 08:43:47.661) (total time: 10518ms): Feb 19 08:43:58 crc kubenswrapper[4675]: Trace[1973479468]: ---"Objects listed" error: 10517ms (08:43:58.178) Feb 19 08:43:58 crc kubenswrapper[4675]: Trace[1973479468]: [10.518054821s] [10.518054821s] END Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.179179 4675 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.179491 4675 trace.go:236] Trace[1690545649]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (19-Feb-2026 08:43:46.908) (total time: 11270ms): Feb 19 08:43:58 crc kubenswrapper[4675]: Trace[1690545649]: ---"Objects listed" error: 11270ms (08:43:58.179) Feb 19 08:43:58 crc kubenswrapper[4675]: Trace[1690545649]: [11.27094628s] [11.27094628s] END Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.179757 4675 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.181439 4675 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.205664 4675 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.232476 4675 csr.go:261] certificate signing request csr-mrf5p is approved, waiting to be issued Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.243333 4675 csr.go:257] certificate signing request csr-mrf5p is issued Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.250111 4675 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46530->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.250196 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46530->192.168.126.11:17697: read: connection reset by peer" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.250150 4675 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46538->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.250728 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46538->192.168.126.11:17697: read: connection reset by peer" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.251348 4675 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.251409 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.251876 4675 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.251963 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.279728 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280157 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280280 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280381 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280470 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280551 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280152 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280633 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280316 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280753 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280790 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280820 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280845 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280871 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280908 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280945 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280968 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280992 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281016 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281039 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281065 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281095 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281129 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281160 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281183 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281209 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281238 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281266 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281291 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281317 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281371 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281400 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281423 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281445 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281467 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281494 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281519 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281544 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281575 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281602 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281628 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281671 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281743 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281768 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281792 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281820 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281846 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281872 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281894 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281920 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280478 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.280842 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281022 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281996 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281208 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281552 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281729 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281928 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.281946 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282560 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282817 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282845 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282872 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282909 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282937 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282964 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.282993 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283023 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283052 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283083 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283109 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283143 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283168 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283196 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283221 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283248 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283243 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283278 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283307 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283331 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283358 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283384 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283406 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283417 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283441 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283467 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283495 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283524 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283553 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283580 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283609 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283653 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283688 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283721 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283747 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283776 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283806 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283832 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283860 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283886 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283910 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283961 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283989 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284018 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284064 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284090 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284121 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284151 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284177 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284201 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284231 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284259 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284285 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284312 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284337 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284366 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284392 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284421 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284449 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284545 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284573 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284599 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284659 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284682 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284709 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284735 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284758 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284786 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284811 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284838 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284861 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284886 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284913 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284936 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284966 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284996 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285022 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285049 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285074 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285102 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285131 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285155 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285184 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285249 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285275 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285299 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285322 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285370 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285397 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285423 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285446 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285472 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285501 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285527 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285550 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285574 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285598 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285621 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285690 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285719 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285750 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285775 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285802 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285829 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285853 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285878 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285902 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285927 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285949 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285973 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285997 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286024 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286051 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286090 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286119 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286145 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286176 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286203 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286231 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286259 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286285 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286316 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286344 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286375 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286403 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286434 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286464 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286493 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286521 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286553 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286583 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286610 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286811 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286849 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286880 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286908 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286937 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286964 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286993 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287028 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287056 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287085 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287118 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287145 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287168 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287189 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287222 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287273 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287308 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287337 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287365 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287387 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287414 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287439 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287462 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287489 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287518 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287543 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287568 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287599 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287627 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287804 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287827 4675 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287840 4675 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287856 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287879 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287894 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287911 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287928 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287942 4675 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287957 4675 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287973 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287987 4675 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283523 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.283714 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284087 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284257 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284531 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284629 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.295924 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284677 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.284917 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285210 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285406 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285534 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285745 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.285794 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286120 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286140 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286199 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286247 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286601 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286733 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286781 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.286839 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287479 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287525 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.287964 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288122 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288078 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288144 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288482 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288508 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288669 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288917 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.288962 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289027 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289198 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289297 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289438 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289561 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289739 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289827 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.289901 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.290768 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.291718 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.291824 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292035 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292146 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292160 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292387 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292420 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292552 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292573 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292602 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292967 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.292998 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.293164 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.293294 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.293447 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.293568 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.293745 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.293863 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.293924 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.294154 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.294720 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.294957 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.296220 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.296226 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.296415 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.296570 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.296688 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.296965 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.297161 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.297365 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.297414 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.297846 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.297963 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.298845 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.299481 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.299706 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.299731 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.299750 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.300251 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.301252 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.300525 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.300694 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.301185 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.301427 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.301454 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.301603 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:58.801467101 +0000 UTC m=+20.428557369 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.301859 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.301536 4675 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.305038 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.301967 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.302118 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.302253 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.302794 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.302844 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.302663 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.303489 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.303473 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.304197 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.304335 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.304478 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.303041 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.304535 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.304592 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.304725 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.305081 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.306501 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.306703 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.306841 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.307553 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.307687 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.307885 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.307997 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.308023 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.308293 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.308813 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.308823 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.305368 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.308930 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.311173 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.311341 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.311967 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.312058 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.312397 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.312393 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.306080 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.312958 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.312990 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.313006 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.313064 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:58.813046684 +0000 UTC m=+20.440137162 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.313727 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.313749 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.305217 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.314183 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.315904 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.316012 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.316227 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.316254 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.316726 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.316813 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.316896 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.317132 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.317193 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.317421 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.317442 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.317486 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:43:58.817454583 +0000 UTC m=+20.444545051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.317768 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.317898 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.318182 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.318307 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.318354 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.318358 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.318533 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.321804 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.323298 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.323391 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.323781 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.323891 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.323924 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.324349 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.324046 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.324605 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.324764 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.324787 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.305460 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.322258 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.322401 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.322828 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.325287 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:58.825247053 +0000 UTC m=+20.452337341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.325875 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.326865 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.322869 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.328206 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.328245 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.329349 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.329603 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.328183 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.328262 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.328261 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.328506 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.328960 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.329220 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.322128 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.330141 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.330179 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.330196 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.330264 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:58.830235258 +0000 UTC m=+20.457325526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.330348 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.330554 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.330761 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.330970 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.330986 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.331113 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.331806 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.331751 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.332337 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.333535 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.333863 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.336440 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.341530 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.351050 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.354350 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.356786 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.362096 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.373580 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-82872935da08b07e21df1e01bcf946881f1bda867e71e214a28cd63c3ed88f4c WatchSource:0}: Error finding container 82872935da08b07e21df1e01bcf946881f1bda867e71e214a28cd63c3ed88f4c: Status 404 returned error can't find the container with id 82872935da08b07e21df1e01bcf946881f1bda867e71e214a28cd63c3ed88f4c Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389320 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389408 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389504 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389554 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389570 4675 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389585 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389600 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389629 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389665 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389680 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389694 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389707 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389738 4675 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389750 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389762 4675 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389773 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389785 4675 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389819 4675 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389833 4675 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389845 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389859 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389871 4675 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389900 4675 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389913 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389925 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389940 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389953 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389983 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.389995 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390009 4675 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390023 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390036 4675 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390066 4675 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390079 4675 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390091 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390105 4675 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390138 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390151 4675 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390167 4675 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390179 4675 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390192 4675 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390226 4675 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390241 4675 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390256 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390268 4675 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390300 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390314 4675 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390327 4675 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390340 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390356 4675 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390386 4675 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390397 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390402 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390479 4675 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390494 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390504 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390516 4675 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390529 4675 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390539 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390549 4675 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390559 4675 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390568 4675 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390579 4675 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390588 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390598 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390608 4675 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390617 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390627 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390657 4675 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390678 4675 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390689 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390698 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390709 4675 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390719 4675 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390729 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390740 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390749 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390758 4675 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390767 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390776 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390788 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390797 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390806 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390816 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390825 4675 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390834 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390843 4675 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390852 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390860 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390879 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390889 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390897 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390907 4675 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390916 4675 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390925 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390935 4675 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390945 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390954 4675 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390967 4675 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390977 4675 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390985 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390996 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391004 4675 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391015 4675 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391023 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391032 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391041 4675 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391051 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391061 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391070 4675 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391080 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391090 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391105 4675 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391174 4675 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391217 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391228 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391802 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391872 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391887 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391907 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391920 4675 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391932 4675 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391944 4675 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391959 4675 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391974 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.391987 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392001 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392013 4675 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392029 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392041 4675 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392054 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392066 4675 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392078 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392091 4675 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392103 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392140 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392480 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392575 4675 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.392598 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393076 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393100 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393113 4675 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393125 4675 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393136 4675 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393148 4675 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393162 4675 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393733 4675 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393762 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393775 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393788 4675 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393801 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393823 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393836 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393849 4675 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393861 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393879 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393892 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393904 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.390192 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.393921 4675 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394088 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394108 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394123 4675 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394136 4675 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394147 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394157 4675 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394169 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394179 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394190 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394205 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394216 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394226 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394235 4675 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394245 4675 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394255 4675 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394267 4675 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394276 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394285 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394294 4675 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394304 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394314 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394323 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394332 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394341 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394350 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.394360 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.650720 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.663965 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.667163 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-7fac0ecb6c391e8cd635b1e521f55fac99619700c087967503ce349f66f27c39 WatchSource:0}: Error finding container 7fac0ecb6c391e8cd635b1e521f55fac99619700c087967503ce349f66f27c39: Status 404 returned error can't find the container with id 7fac0ecb6c391e8cd635b1e521f55fac99619700c087967503ce349f66f27c39 Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.863796 4675 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.864026 4675 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events\": read tcp 38.102.83.158:46968->38.102.83.158:6443: use of closed network connection" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189599552f666b61 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-19 08:43:39.586669409 +0000 UTC m=+1.213759677,LastTimestamp:2026-02-19 08:43:39.586669409 +0000 UTC m=+1.213759677,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864149 4675 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864192 4675 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864196 4675 reflector.go:484] pkg/kubelet/config/apiserver.go:66: watch of *v1.Pod ended with: very short watch: pkg/kubelet/config/apiserver.go:66: Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864218 4675 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864187 4675 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864226 4675 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864256 4675 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864260 4675 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864255 4675 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864278 4675 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864241 4675 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864312 4675 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: W0219 08:43:58.864348 4675 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.899078 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.899151 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.899176 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.899193 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:43:58 crc kubenswrapper[4675]: I0219 08:43:58.899210 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899315 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899329 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899339 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899394 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:59.899381371 +0000 UTC m=+21.526471639 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899441 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:43:59.899436382 +0000 UTC m=+21.526526650 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899483 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899502 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:59.899497224 +0000 UTC m=+21.526587492 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899539 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899549 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899557 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899576 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:59.899569516 +0000 UTC m=+21.526659794 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899605 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:43:58 crc kubenswrapper[4675]: E0219 08:43:58.899624 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:43:59.899618857 +0000 UTC m=+21.526709125 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.049619 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 23:57:12.877714359 +0000 UTC Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.102895 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.103038 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.106513 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.107045 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.107819 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.108403 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.108952 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.109442 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.110017 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.110533 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.111184 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.113210 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.113926 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.114948 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.115597 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.116262 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.116924 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.117551 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.118259 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.118742 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.119382 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.119561 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.120097 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.120630 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.121335 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.121892 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.122673 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.123144 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.123862 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.124592 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.126263 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.126978 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.127598 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.128259 4675 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.128383 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.130118 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.130661 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.131156 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.132504 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.133223 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.133824 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.134440 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.134686 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.135214 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.135746 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.139462 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.140279 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.141329 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.141882 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.143162 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.143900 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.145131 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.145598 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.146470 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.146942 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.147829 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.148376 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.148848 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.153082 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.167570 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.183327 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.197238 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.242201 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.245192 4675 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-19 08:38:58 +0000 UTC, rotation deadline is 2026-12-31 02:20:06.510447339 +0000 UTC Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.245359 4675 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7553h36m7.265092338s for next certificate rotation Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.246053 4675 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0" exitCode=255 Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.246241 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0"} Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.249994 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d03d2b3dcc5e02a761588d63d48178787a2f6cbe164b35c5dfaca3f54f363863"} Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.251898 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df"} Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.251926 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7fac0ecb6c391e8cd635b1e521f55fac99619700c087967503ce349f66f27c39"} Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.254340 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e"} Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.254380 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db"} Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.254391 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"82872935da08b07e21df1e01bcf946881f1bda867e71e214a28cd63c3ed88f4c"} Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.260587 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.280065 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.301676 4675 scope.go:117] "RemoveContainer" containerID="456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.302004 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.330609 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.375423 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.438074 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.458021 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.489223 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.514139 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.558425 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.596320 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.651043 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.777450 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.872481 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.905039 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.905161 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.905202 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.905237 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.905266 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905448 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905478 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905495 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905562 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:01.905541044 +0000 UTC m=+23.532631322 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905658 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:44:01.905631326 +0000 UTC m=+23.532721604 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905712 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905742 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:01.905734149 +0000 UTC m=+23.532824417 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905808 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905822 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905832 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905893 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:01.905884593 +0000 UTC m=+23.532974871 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.905926 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: E0219 08:43:59.906024 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:01.906004057 +0000 UTC m=+23.533094315 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.958853 4675 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.959547 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hj427","openshift-machine-config-operator/machine-config-daemon-wqmkn","openshift-image-registry/node-ca-6kmw9","openshift-kube-apiserver/kube-apiserver-crc"] Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.960207 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hj427" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.960372 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.960432 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.961994 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.962132 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.963364 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965352 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965482 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965578 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965611 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965704 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965823 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965850 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.965952 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.966039 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 19 08:43:59 crc kubenswrapper[4675]: I0219 08:43:59.987698 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:43:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.002459 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.017665 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.032478 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.046079 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.050422 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 23:46:03.772089226 +0000 UTC Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.057910 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.073925 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.101851 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.103613 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.103719 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:00 crc kubenswrapper[4675]: E0219 08:44:00.103765 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:00 crc kubenswrapper[4675]: E0219 08:44:00.103913 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106608 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c92c6f02-3c18-4d7a-b559-840c8415db58-hosts-file\") pod \"node-resolver-hj427\" (UID: \"c92c6f02-3c18-4d7a-b559-840c8415db58\") " pod="openshift-dns/node-resolver-hj427" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106722 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjb8q\" (UniqueName: \"kubernetes.io/projected/c92c6f02-3c18-4d7a-b559-840c8415db58-kube-api-access-cjb8q\") pod \"node-resolver-hj427\" (UID: \"c92c6f02-3c18-4d7a-b559-840c8415db58\") " pod="openshift-dns/node-resolver-hj427" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106776 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4j22\" (UniqueName: \"kubernetes.io/projected/5e91469d-12b7-434c-991d-633287712a69-kube-api-access-k4j22\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106820 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e91469d-12b7-434c-991d-633287712a69-mcd-auth-proxy-config\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106853 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2frs\" (UniqueName: \"kubernetes.io/projected/5c50286d-89aa-4d1c-a424-5fd45d0355fb-kube-api-access-k2frs\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106894 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5e91469d-12b7-434c-991d-633287712a69-rootfs\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106928 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5c50286d-89aa-4d1c-a424-5fd45d0355fb-serviceca\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.106962 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c50286d-89aa-4d1c-a424-5fd45d0355fb-host\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.107034 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e91469d-12b7-434c-991d-633287712a69-proxy-tls\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.113792 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.117973 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.119061 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.129422 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.137871 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.155795 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-qmz2s"] Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.156236 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-h6jd6"] Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.156390 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.156532 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.157208 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.158867 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.159167 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.159841 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.159851 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.160265 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.163846 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.163898 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.176324 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.188321 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.200437 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.205912 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207408 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5e91469d-12b7-434c-991d-633287712a69-rootfs\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207445 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5c50286d-89aa-4d1c-a424-5fd45d0355fb-serviceca\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207478 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-cni-bin\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207507 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-cnibin\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207531 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-hostroot\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207550 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5e91469d-12b7-434c-991d-633287712a69-rootfs\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207557 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-conf-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207676 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c92c6f02-3c18-4d7a-b559-840c8415db58-hosts-file\") pod \"node-resolver-hj427\" (UID: \"c92c6f02-3c18-4d7a-b559-840c8415db58\") " pod="openshift-dns/node-resolver-hj427" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207710 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-socket-dir-parent\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207763 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c92c6f02-3c18-4d7a-b559-840c8415db58-hosts-file\") pod \"node-resolver-hj427\" (UID: \"c92c6f02-3c18-4d7a-b559-840c8415db58\") " pod="openshift-dns/node-resolver-hj427" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207826 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e91469d-12b7-434c-991d-633287712a69-mcd-auth-proxy-config\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207891 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-system-cni-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207916 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-daemon-config\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207938 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-etc-kubernetes\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207960 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-multus-certs\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.207978 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzdkh\" (UniqueName: \"kubernetes.io/projected/4cc07c5a-23e4-478f-88a7-73d41e17a320-kube-api-access-tzdkh\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208010 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c50286d-89aa-4d1c-a424-5fd45d0355fb-host\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208031 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-netns\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208069 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e91469d-12b7-434c-991d-633287712a69-proxy-tls\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208072 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c50286d-89aa-4d1c-a424-5fd45d0355fb-host\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208089 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-os-release\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208118 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjb8q\" (UniqueName: \"kubernetes.io/projected/c92c6f02-3c18-4d7a-b559-840c8415db58-kube-api-access-cjb8q\") pod \"node-resolver-hj427\" (UID: \"c92c6f02-3c18-4d7a-b559-840c8415db58\") " pod="openshift-dns/node-resolver-hj427" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208134 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-cni-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208150 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4cc07c5a-23e4-478f-88a7-73d41e17a320-cni-binary-copy\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208167 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-kubelet\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208197 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4j22\" (UniqueName: \"kubernetes.io/projected/5e91469d-12b7-434c-991d-633287712a69-kube-api-access-k4j22\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208319 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-k8s-cni-cncf-io\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208366 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2frs\" (UniqueName: \"kubernetes.io/projected/5c50286d-89aa-4d1c-a424-5fd45d0355fb-kube-api-access-k2frs\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208398 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-cni-multus\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208543 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5c50286d-89aa-4d1c-a424-5fd45d0355fb-serviceca\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.208790 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e91469d-12b7-434c-991d-633287712a69-mcd-auth-proxy-config\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.215386 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e91469d-12b7-434c-991d-633287712a69-proxy-tls\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.220573 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.224242 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjb8q\" (UniqueName: \"kubernetes.io/projected/c92c6f02-3c18-4d7a-b559-840c8415db58-kube-api-access-cjb8q\") pod \"node-resolver-hj427\" (UID: \"c92c6f02-3c18-4d7a-b559-840c8415db58\") " pod="openshift-dns/node-resolver-hj427" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.228232 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4j22\" (UniqueName: \"kubernetes.io/projected/5e91469d-12b7-434c-991d-633287712a69-kube-api-access-k4j22\") pod \"machine-config-daemon-wqmkn\" (UID: \"5e91469d-12b7-434c-991d-633287712a69\") " pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.229084 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2frs\" (UniqueName: \"kubernetes.io/projected/5c50286d-89aa-4d1c-a424-5fd45d0355fb-kube-api-access-k2frs\") pod \"node-ca-6kmw9\" (UID: \"5c50286d-89aa-4d1c-a424-5fd45d0355fb\") " pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.237290 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.240604 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.254841 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.258426 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.260786 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced"} Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.261013 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:44:00 crc kubenswrapper[4675]: E0219 08:44:00.268541 4675 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.269161 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.270334 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.273619 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hj427" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.280081 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.284083 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.285657 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6kmw9" Feb 19 08:44:00 crc kubenswrapper[4675]: W0219 08:44:00.286203 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc92c6f02_3c18_4d7a_b559_840c8415db58.slice/crio-d53c3e322b8823b68dde8fd3784459622b7e2415d7c776ea0cfc7b04fe6d5a94 WatchSource:0}: Error finding container d53c3e322b8823b68dde8fd3784459622b7e2415d7c776ea0cfc7b04fe6d5a94: Status 404 returned error can't find the container with id d53c3e322b8823b68dde8fd3784459622b7e2415d7c776ea0cfc7b04fe6d5a94 Feb 19 08:44:00 crc kubenswrapper[4675]: W0219 08:44:00.296070 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e91469d_12b7_434c_991d_633287712a69.slice/crio-e82683792b6cefca88629560848c2771f723626f9e4cacd8a868bf011ff3cf69 WatchSource:0}: Error finding container e82683792b6cefca88629560848c2771f723626f9e4cacd8a868bf011ff3cf69: Status 404 returned error can't find the container with id e82683792b6cefca88629560848c2771f723626f9e4cacd8a868bf011ff3cf69 Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.298075 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: W0219 08:44:00.304451 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c50286d_89aa_4d1c_a424_5fd45d0355fb.slice/crio-c74107d337823c94dc6aa88ac4f66c4f4c30088812e65a5285e0f2ca799040bb WatchSource:0}: Error finding container c74107d337823c94dc6aa88ac4f66c4f4c30088812e65a5285e0f2ca799040bb: Status 404 returned error can't find the container with id c74107d337823c94dc6aa88ac4f66c4f4c30088812e65a5285e0f2ca799040bb Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309019 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-cnibin\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309064 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-hostroot\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309105 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-conf-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309132 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-socket-dir-parent\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309171 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309202 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7912463c-4c5e-4269-a911-814275e128b0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309229 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-system-cni-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309241 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-cnibin\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309257 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-daemon-config\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309358 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-etc-kubernetes\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309385 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-multus-certs\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309404 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzdkh\" (UniqueName: \"kubernetes.io/projected/4cc07c5a-23e4-478f-88a7-73d41e17a320-kube-api-access-tzdkh\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309436 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-cnibin\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309470 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-netns\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309499 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-system-cni-dir\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309553 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-os-release\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309603 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-cni-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309623 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4cc07c5a-23e4-478f-88a7-73d41e17a320-cni-binary-copy\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309675 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-kubelet\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309698 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-k8s-cni-cncf-io\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309727 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-os-release\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309761 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrmjm\" (UniqueName: \"kubernetes.io/projected/7912463c-4c5e-4269-a911-814275e128b0-kube-api-access-qrmjm\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309783 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-cni-multus\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309809 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7912463c-4c5e-4269-a911-814275e128b0-cni-binary-copy\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309844 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-cni-bin\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309908 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-cni-bin\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309935 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-etc-kubernetes\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.309961 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-multus-certs\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310012 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-daemon-config\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310068 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-hostroot\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310093 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-conf-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310143 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-socket-dir-parent\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310202 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-system-cni-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310327 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-kubelet\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310359 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-k8s-cni-cncf-io\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310400 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-var-lib-cni-multus\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310443 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-host-run-netns\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310632 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4cc07c5a-23e4-478f-88a7-73d41e17a320-cni-binary-copy\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.310714 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-os-release\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.311016 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4cc07c5a-23e4-478f-88a7-73d41e17a320-multus-cni-dir\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.312148 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.331560 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.343972 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.345959 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzdkh\" (UniqueName: \"kubernetes.io/projected/4cc07c5a-23e4-478f-88a7-73d41e17a320-kube-api-access-tzdkh\") pod \"multus-qmz2s\" (UID: \"4cc07c5a-23e4-478f-88a7-73d41e17a320\") " pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.366887 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.388937 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.394550 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.413295 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.413356 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7912463c-4c5e-4269-a911-814275e128b0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.413397 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-cnibin\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.413423 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-system-cni-dir\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.413474 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-os-release\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.413513 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrmjm\" (UniqueName: \"kubernetes.io/projected/7912463c-4c5e-4269-a911-814275e128b0-kube-api-access-qrmjm\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.413539 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7912463c-4c5e-4269-a911-814275e128b0-cni-binary-copy\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.414632 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7912463c-4c5e-4269-a911-814275e128b0-cni-binary-copy\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.414706 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.414768 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-cnibin\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.414735 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-system-cni-dir\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.414848 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7912463c-4c5e-4269-a911-814275e128b0-os-release\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.415086 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7912463c-4c5e-4269-a911-814275e128b0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.416782 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.427041 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.444346 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.450004 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrmjm\" (UniqueName: \"kubernetes.io/projected/7912463c-4c5e-4269-a911-814275e128b0-kube-api-access-qrmjm\") pod \"multus-additional-cni-plugins-h6jd6\" (UID: \"7912463c-4c5e-4269-a911-814275e128b0\") " pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.457751 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.474270 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.476762 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qmz2s" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.482941 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.505042 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.523064 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.607951 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lkvd2"] Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.608999 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.639524 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.639741 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.639821 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.639887 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.639914 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.639916 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.647479 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.718208 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719184 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-netns\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719237 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-ovn-kubernetes\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719276 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-etc-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719293 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5ndr\" (UniqueName: \"kubernetes.io/projected/d95acb1e-cd60-48dd-af45-e190cd07723a-kube-api-access-z5ndr\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719310 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719327 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-netd\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719349 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-script-lib\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719364 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-node-log\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719382 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-kubelet\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719395 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-systemd-units\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719412 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-slash\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719432 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-systemd\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719458 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719482 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d95acb1e-cd60-48dd-af45-e190cd07723a-ovn-node-metrics-cert\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719499 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-var-lib-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719529 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-config\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719546 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-log-socket\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719564 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-bin\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719591 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-env-overrides\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.719620 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-ovn\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.803966 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821603 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-var-lib-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821687 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-config\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821708 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-bin\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821733 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-log-socket\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821766 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-env-overrides\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821800 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-ovn\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821804 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-var-lib-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821888 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-netns\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821833 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-netns\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821962 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-ovn-kubernetes\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821819 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-bin\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.821998 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-etc-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822016 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-ovn\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822039 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-ovn-kubernetes\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822026 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5ndr\" (UniqueName: \"kubernetes.io/projected/d95acb1e-cd60-48dd-af45-e190cd07723a-kube-api-access-z5ndr\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822086 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-log-socket\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822139 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-netd\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822099 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-etc-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822111 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-netd\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822245 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822273 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-script-lib\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822282 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-openvswitch\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822338 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-kubelet\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822302 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-kubelet\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822405 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-systemd-units\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822432 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-node-log\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822442 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-systemd-units\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822466 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-slash\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822488 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-node-log\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822506 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d95acb1e-cd60-48dd-af45-e190cd07723a-ovn-node-metrics-cert\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822526 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-slash\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822546 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-systemd\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822577 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822688 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-systemd\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822714 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.822811 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-env-overrides\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.823012 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-config\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.823691 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-script-lib\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.827215 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d95acb1e-cd60-48dd-af45-e190cd07723a-ovn-node-metrics-cert\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.865667 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5ndr\" (UniqueName: \"kubernetes.io/projected/d95acb1e-cd60-48dd-af45-e190cd07723a-kube-api-access-z5ndr\") pod \"ovnkube-node-lkvd2\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.871194 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.895733 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.914420 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.927265 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.937237 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.952004 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.966364 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.978265 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:00 crc kubenswrapper[4675]: I0219 08:44:00.993903 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:00Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.009799 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.024355 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.038942 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.051477 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 23:06:41.831913505 +0000 UTC Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.054128 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.067044 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.082880 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.096808 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.102426 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.102919 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.130303 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.145544 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.265033 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerStarted","Data":"05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.265122 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerStarted","Data":"3ed08d5488b20bc384bac49206ac688f086abab05b1b8eece58c3bebe0bdf617"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.268078 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6kmw9" event={"ID":"5c50286d-89aa-4d1c-a424-5fd45d0355fb","Type":"ContainerStarted","Data":"74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.268143 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6kmw9" event={"ID":"5c50286d-89aa-4d1c-a424-5fd45d0355fb","Type":"ContainerStarted","Data":"c74107d337823c94dc6aa88ac4f66c4f4c30088812e65a5285e0f2ca799040bb"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.270456 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.270497 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.270509 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"e82683792b6cefca88629560848c2771f723626f9e4cacd8a868bf011ff3cf69"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.273360 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hj427" event={"ID":"c92c6f02-3c18-4d7a-b559-840c8415db58","Type":"ContainerStarted","Data":"44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.273442 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hj427" event={"ID":"c92c6f02-3c18-4d7a-b559-840c8415db58","Type":"ContainerStarted","Data":"d53c3e322b8823b68dde8fd3784459622b7e2415d7c776ea0cfc7b04fe6d5a94"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.275344 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.276718 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerStarted","Data":"b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.276876 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerStarted","Data":"17ebb04aec7d3e5d314171fe5111a2607d6faff1a6e32935f1591e2ca125d86a"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.277946 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae" exitCode=0 Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.278537 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.278576 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"805386fab984c5af04d7bf2953d3e47915f49059662b00a9cdde597229de5a30"} Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.290471 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.323590 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.337001 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.353328 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.366149 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.381227 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.404592 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.419692 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.434712 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.450631 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.464784 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.481710 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.496411 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.523288 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.538448 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.562772 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.579689 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.593267 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.613393 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.630680 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.643244 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.658627 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.677977 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.693006 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.725172 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.763917 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.809745 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.846730 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:01Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.932784 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.932986 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.933025 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933126 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:44:05.933085132 +0000 UTC m=+27.560175400 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933163 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933146 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933191 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933208 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933255 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:05.933242176 +0000 UTC m=+27.560332444 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.933306 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933358 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:05.933342729 +0000 UTC m=+27.560433217 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:01 crc kubenswrapper[4675]: I0219 08:44:01.933399 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933517 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933557 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933561 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933587 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933606 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:05.933595776 +0000 UTC m=+27.560686044 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:01 crc kubenswrapper[4675]: E0219 08:44:01.933694 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:05.933662088 +0000 UTC m=+27.560752386 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.052626 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 04:15:56.270278208 +0000 UTC Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.103204 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.103262 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:02 crc kubenswrapper[4675]: E0219 08:44:02.103385 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:02 crc kubenswrapper[4675]: E0219 08:44:02.103488 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.169437 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.185277 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.188182 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.188261 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.203236 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.217926 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.236378 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.249799 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.265099 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.289386 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.290833 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e"} Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.290889 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a"} Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.290903 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb"} Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.290916 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830"} Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.290928 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d"} Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.290939 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a"} Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.295245 4675 generic.go:334] "Generic (PLEG): container finished" podID="7912463c-4c5e-4269-a911-814275e128b0" containerID="05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308" exitCode=0 Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.296187 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerDied","Data":"05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308"} Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.307347 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.320918 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.338836 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.353610 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.363119 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.381846 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.423681 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.461973 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.509809 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.547649 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.589016 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.622848 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.665054 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.715758 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.752455 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.797771 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.825016 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.870697 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.906452 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:02 crc kubenswrapper[4675]: I0219 08:44:02.943996 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.034075 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:02Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.054360 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 21:11:39.224833481 +0000 UTC Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.062573 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.103693 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:03 crc kubenswrapper[4675]: E0219 08:44:03.103872 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.300385 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerStarted","Data":"2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225"} Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.319206 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.337215 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.350756 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.361434 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.377068 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.393020 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.407983 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.427318 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.441896 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.465905 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.482208 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.505132 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.544154 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.590047 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:03 crc kubenswrapper[4675]: I0219 08:44:03.622708 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:03Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.056045 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 06:41:56.341740123 +0000 UTC Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.107506 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.107684 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.107775 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.107844 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.307891 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0"} Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.311202 4675 generic.go:334] "Generic (PLEG): container finished" podID="7912463c-4c5e-4269-a911-814275e128b0" containerID="2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225" exitCode=0 Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.311250 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerDied","Data":"2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225"} Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.331791 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.349609 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.366487 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.379570 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.397340 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.413222 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.438560 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.456174 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.473208 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.512730 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.530569 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.546566 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.564937 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.581583 4675 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.584299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.584346 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.584359 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.584541 4675 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.594305 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.602056 4675 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.602428 4675 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.603992 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.604012 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.604021 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.604041 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.604053 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.613927 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.623482 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.628551 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.628610 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.628697 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.628758 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.628771 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.647020 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.653064 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.653113 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.653126 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.653148 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.653161 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.667463 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.672141 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.672192 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.672208 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.672232 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.672245 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.688086 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.693286 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.693338 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.693353 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.693375 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.693387 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.707519 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:04Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:04 crc kubenswrapper[4675]: E0219 08:44:04.707756 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.712149 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.712530 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.712610 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.712776 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.712889 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.815712 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.815756 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.815769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.815787 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.815798 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.918768 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.918829 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.918843 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.918871 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:04 crc kubenswrapper[4675]: I0219 08:44:04.918897 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:04Z","lastTransitionTime":"2026-02-19T08:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.022763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.022813 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.022828 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.022849 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.022862 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.056928 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 16:01:18.689949211 +0000 UTC Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.103158 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.103342 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.131544 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.131587 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.131599 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.131618 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.131633 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.234821 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.234866 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.234876 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.234899 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.234909 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.317710 4675 generic.go:334] "Generic (PLEG): container finished" podID="7912463c-4c5e-4269-a911-814275e128b0" containerID="2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3" exitCode=0 Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.317794 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerDied","Data":"2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.338316 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.343831 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.343879 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.343902 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.343923 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.343936 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.356836 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.370920 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.386082 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.398770 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.414598 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.429519 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.444793 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.447971 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.448021 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.448034 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.448063 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.448092 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.462893 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.482817 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.502010 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.515965 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.530095 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.550171 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.551549 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.551596 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.551689 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.551726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.551768 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.563942 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:05Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.655245 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.655339 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.655356 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.655385 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.655403 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.758886 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.758930 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.758939 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.758958 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.758970 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.861393 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.861436 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.861448 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.861470 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.861483 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.964686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.964741 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.964751 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.964771 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.964783 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:05Z","lastTransitionTime":"2026-02-19T08:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.978386 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.978562 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.978632 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.978739 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:44:13.978690962 +0000 UTC m=+35.605781230 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.978773 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.978828 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.978876 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:13.978850576 +0000 UTC m=+35.605940844 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.978797 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.978904 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: I0219 08:44:05.978900 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.978920 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.978969 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:13.978956939 +0000 UTC m=+35.606047207 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.979022 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.979026 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.979055 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.979071 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.979057 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:13.979049332 +0000 UTC m=+35.606139590 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:05 crc kubenswrapper[4675]: E0219 08:44:05.979164 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:13.979132184 +0000 UTC m=+35.606222452 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.058897 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 04:44:10.122383196 +0000 UTC Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.067292 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.067347 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.067363 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.067387 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.067403 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.102781 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.102850 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:06 crc kubenswrapper[4675]: E0219 08:44:06.103006 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:06 crc kubenswrapper[4675]: E0219 08:44:06.103107 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.169390 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.169440 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.169451 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.169468 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.169481 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.272302 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.272351 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.272360 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.272377 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.272387 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.324856 4675 generic.go:334] "Generic (PLEG): container finished" podID="7912463c-4c5e-4269-a911-814275e128b0" containerID="c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7" exitCode=0 Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.324920 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerDied","Data":"c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.347230 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.362019 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.376609 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.376684 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.376699 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.376722 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.376734 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.377316 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.396841 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.410607 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.426878 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.450750 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.467448 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.480118 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.480154 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.480163 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.480181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.480191 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.493035 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.505902 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.523179 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.536560 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.551773 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.572892 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.586667 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.588063 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.588081 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.588089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.588105 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.588114 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.692278 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.692737 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.692747 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.692766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.692777 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.775918 4675 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.801956 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.802026 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.802040 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.802061 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.802075 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.905109 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.905147 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.905160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.905181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:06 crc kubenswrapper[4675]: I0219 08:44:06.905195 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:06Z","lastTransitionTime":"2026-02-19T08:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.015341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.015394 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.015408 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.015432 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.015442 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.063784 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 03:31:10.618417192 +0000 UTC Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.107846 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:07 crc kubenswrapper[4675]: E0219 08:44:07.108031 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.118299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.118343 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.118355 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.118473 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.118500 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.147722 4675 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.221804 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.221873 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.221888 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.221913 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.221929 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.324971 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.325009 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.325020 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.325039 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.325060 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.332532 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.333551 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.333612 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.337424 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerStarted","Data":"4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.347568 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.365576 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.372838 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.373541 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.383723 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.398265 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.415398 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.427834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.427901 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.427916 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.427968 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.427984 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.432879 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.455176 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.472796 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.489406 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.505901 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.521136 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.531246 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.531293 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.531303 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.531320 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.531332 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.536227 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.551607 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.572398 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.588170 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.605828 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.618371 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.632002 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.633428 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.633459 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.633471 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.633490 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.633501 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.646883 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.664128 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.677625 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.691189 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.713146 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.733241 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.736298 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.736330 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.736339 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.736357 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.736368 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.748786 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.764236 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.779042 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.791991 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.809970 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.824929 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:07Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.839335 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.839385 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.839400 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.839426 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.839446 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.913169 4675 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.942731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.942800 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.942816 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.942846 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:07 crc kubenswrapper[4675]: I0219 08:44:07.942860 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:07Z","lastTransitionTime":"2026-02-19T08:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.045273 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.045316 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.045324 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.045339 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.045349 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.064963 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 20:26:40.90894882 +0000 UTC Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.102309 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.102378 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:08 crc kubenswrapper[4675]: E0219 08:44:08.102478 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:08 crc kubenswrapper[4675]: E0219 08:44:08.102580 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.147773 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.147816 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.147827 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.147847 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.147861 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.250829 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.250891 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.250907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.250935 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.250952 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.345236 4675 generic.go:334] "Generic (PLEG): container finished" podID="7912463c-4c5e-4269-a911-814275e128b0" containerID="4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21" exitCode=0 Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.345334 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerDied","Data":"4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.345419 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.356490 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.357342 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.357458 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.357537 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.357605 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.375168 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.402289 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.417482 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.428665 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.443207 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.457604 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.463212 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.463264 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.463278 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.463301 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.463314 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.473102 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.489332 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.503875 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.528178 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.547060 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.562537 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.566318 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.566348 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.566360 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.566376 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.566386 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.577873 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.597399 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.611456 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:08Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.668948 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.668996 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.669009 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.669028 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.669041 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.771952 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.772010 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.772027 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.772053 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.772072 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.876400 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.876849 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.876936 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.877031 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.877100 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.981282 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.981707 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.981800 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.981884 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:08 crc kubenswrapper[4675]: I0219 08:44:08.981958 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:08Z","lastTransitionTime":"2026-02-19T08:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.065750 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 08:34:54.836039664 +0000 UTC Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.085685 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.085753 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.085769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.085789 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.085799 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.103097 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:09 crc kubenswrapper[4675]: E0219 08:44:09.103671 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.128786 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.145845 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.164703 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.182858 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.192010 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.192067 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.192079 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.192100 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.192112 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.201287 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.217936 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.235813 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.263026 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.289899 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.295691 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.295758 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.295777 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.295804 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.295819 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.313366 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.332083 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.343604 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.352929 4675 generic.go:334] "Generic (PLEG): container finished" podID="7912463c-4c5e-4269-a911-814275e128b0" containerID="452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb" exitCode=0 Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.353112 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.354385 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerDied","Data":"452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.367112 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.383524 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.399616 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.399736 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.399749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.399770 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.399786 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.403857 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.424917 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.439614 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.451242 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.465551 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.482592 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.499555 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.502570 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.502612 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.502621 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.502663 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.502682 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.540019 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.558072 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.583486 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.606293 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.606371 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.606385 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.606410 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.606438 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.611352 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.633232 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.648921 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.666168 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.694429 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.711612 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.711685 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.711699 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.711722 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.711736 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.712224 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.814151 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.814188 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.814199 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.814213 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.814223 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.917387 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.917443 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.917454 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.917482 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:09 crc kubenswrapper[4675]: I0219 08:44:09.917519 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:09Z","lastTransitionTime":"2026-02-19T08:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.021800 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.021849 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.021861 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.021890 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.021903 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.066201 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 06:20:52.912770906 +0000 UTC Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.102574 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.102601 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:10 crc kubenswrapper[4675]: E0219 08:44:10.102744 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:10 crc kubenswrapper[4675]: E0219 08:44:10.102937 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.124455 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.124498 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.124509 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.124529 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.124540 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.227436 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.227479 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.227488 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.227508 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.227520 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.331373 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.331434 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.331447 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.331467 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.331480 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.359358 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/0.log" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.362168 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23" exitCode=1 Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.362268 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.363284 4675 scope.go:117] "RemoveContainer" containerID="e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.366949 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" event={"ID":"7912463c-4c5e-4269-a911-814275e128b0","Type":"ContainerStarted","Data":"64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.388835 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.403595 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.421186 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.435280 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.435341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.435358 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.435385 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.435400 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.436368 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.456593 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.472478 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.487539 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.502092 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.524265 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.540475 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.540800 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.540819 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.540829 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.540847 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.540860 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.556713 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.575613 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.591805 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.603998 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.617042 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.632404 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.644429 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.644470 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.644478 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.644495 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.644507 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.646525 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.659944 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.673226 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.689767 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.705463 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.729050 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.747917 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.747959 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.747970 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.747990 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.748005 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.751360 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.776414 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.791890 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.806401 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.821068 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.849131 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.850804 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.850865 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.850884 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.850910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.850925 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.866394 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.900028 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:10Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.958769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.958831 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.958845 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.958869 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:10 crc kubenswrapper[4675]: I0219 08:44:10.958882 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:10Z","lastTransitionTime":"2026-02-19T08:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.066800 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 13:27:31.861672637 +0000 UTC Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.088130 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.088173 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.088183 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.088207 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.088217 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.102623 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:11 crc kubenswrapper[4675]: E0219 08:44:11.102826 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.191751 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.191827 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.191837 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.191857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.191870 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.294600 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.294662 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.294672 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.294696 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.294708 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.372626 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/0.log" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.375041 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.375179 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.390895 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.397146 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.397194 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.397207 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.397228 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.397240 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.408688 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.434077 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.450465 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.464118 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.478676 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.490447 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.500251 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.500326 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.500339 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.500389 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.500409 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.504038 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.531165 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.544208 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.559833 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.573779 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.587986 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.603695 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.604304 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.604384 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.604450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.604516 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.607605 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.626406 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.707624 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.707967 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.708049 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.708134 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.708218 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.811592 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.811663 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.811673 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.811691 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.811700 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.915659 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.915727 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.915739 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.915759 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:11 crc kubenswrapper[4675]: I0219 08:44:11.915773 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:11Z","lastTransitionTime":"2026-02-19T08:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.018701 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.018754 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.018767 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.018786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.018798 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.067005 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 23:35:02.117089787 +0000 UTC Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.103116 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.103172 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:12 crc kubenswrapper[4675]: E0219 08:44:12.103265 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:12 crc kubenswrapper[4675]: E0219 08:44:12.103332 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.121808 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.121859 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.121870 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.121892 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.121903 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.225873 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.225942 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.225956 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.225980 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.225994 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.329058 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.329120 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.329131 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.329150 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.329161 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.380899 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/1.log" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.381413 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/0.log" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.383274 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0" exitCode=1 Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.383339 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.383475 4675 scope.go:117] "RemoveContainer" containerID="e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.384274 4675 scope.go:117] "RemoveContainer" containerID="d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0" Feb 19 08:44:12 crc kubenswrapper[4675]: E0219 08:44:12.384465 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.405076 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.418602 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.431174 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.432496 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.432548 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.432563 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.432584 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.432595 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.445417 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.462262 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.482385 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.497710 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.515188 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.529461 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.535596 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.535671 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.535688 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.535714 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.535727 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.544716 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.560861 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.575415 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.595691 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.613609 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.624874 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:12Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.639426 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.639499 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.639511 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.639528 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.639562 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.743589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.743997 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.744074 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.744198 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.744268 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.847106 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.847167 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.847177 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.847198 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.847211 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.950675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.950748 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.950761 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.950784 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.950793 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:12Z","lastTransitionTime":"2026-02-19T08:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.987988 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85"] Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.988585 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.995556 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 19 08:44:12 crc kubenswrapper[4675]: I0219 08:44:12.995827 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.008491 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.025450 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.054065 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.054142 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.054167 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.054198 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.054214 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.054270 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.067567 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 09:45:21.399401333 +0000 UTC Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.073077 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.089218 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.103033 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:13 crc kubenswrapper[4675]: E0219 08:44:13.103256 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.106744 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.123577 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.138623 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.156491 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.156563 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.156580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.156609 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.156629 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.159923 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.169036 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/18ae2562-beb3-4423-9045-f8559d492d8a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.169091 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj8nk\" (UniqueName: \"kubernetes.io/projected/18ae2562-beb3-4423-9045-f8559d492d8a-kube-api-access-tj8nk\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.169179 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/18ae2562-beb3-4423-9045-f8559d492d8a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.169254 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/18ae2562-beb3-4423-9045-f8559d492d8a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.172247 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.186524 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.201337 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.215417 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.231855 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.247800 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.260089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.260149 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.260160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.260181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.260201 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.264053 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:13Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.270029 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/18ae2562-beb3-4423-9045-f8559d492d8a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.270091 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/18ae2562-beb3-4423-9045-f8559d492d8a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.270120 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/18ae2562-beb3-4423-9045-f8559d492d8a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.270139 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj8nk\" (UniqueName: \"kubernetes.io/projected/18ae2562-beb3-4423-9045-f8559d492d8a-kube-api-access-tj8nk\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.270999 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/18ae2562-beb3-4423-9045-f8559d492d8a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.271367 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/18ae2562-beb3-4423-9045-f8559d492d8a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.278226 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/18ae2562-beb3-4423-9045-f8559d492d8a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.289509 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj8nk\" (UniqueName: \"kubernetes.io/projected/18ae2562-beb3-4423-9045-f8559d492d8a-kube-api-access-tj8nk\") pod \"ovnkube-control-plane-749d76644c-x2w85\" (UID: \"18ae2562-beb3-4423-9045-f8559d492d8a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.307323 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.367453 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.367503 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.367516 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.367535 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.367546 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.387853 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" event={"ID":"18ae2562-beb3-4423-9045-f8559d492d8a","Type":"ContainerStarted","Data":"165070689f8b1b810275b84460dd79551d608e635498eaa5848be00564639431"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.394845 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/1.log" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.470593 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.470663 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.470673 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.470693 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.470713 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.572574 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.572616 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.572626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.572664 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.572682 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.675925 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.675971 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.675982 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.676003 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.676015 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.778655 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.778714 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.778722 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.778741 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.778754 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.881049 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.881113 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.881129 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.881146 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.881158 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.983988 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.984051 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.984062 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.984083 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:13 crc kubenswrapper[4675]: I0219 08:44:13.984094 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:13Z","lastTransitionTime":"2026-02-19T08:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.067735 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 18:45:07.954926883 +0000 UTC Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.078421 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.078544 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.078583 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.078607 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.078633 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078724 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078768 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078787 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:30.078767772 +0000 UTC m=+51.705858040 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078787 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078807 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078842 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:30.078830783 +0000 UTC m=+51.705921041 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078881 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:44:30.078871894 +0000 UTC m=+51.705962162 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078877 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.078987 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:30.078960337 +0000 UTC m=+51.706050655 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.079001 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.079052 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.079068 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.079146 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:44:30.079122191 +0000 UTC m=+51.706212459 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.087279 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.087328 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.087342 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.087363 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.087377 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.099420 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vcqbz"] Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.100078 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.100162 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.102409 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.102409 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.102504 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.102581 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.122430 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.136551 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.150762 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.164847 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.177142 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.189684 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.189731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.189742 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.189765 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.189779 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.193387 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.206774 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.220384 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.237893 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.256509 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.279463 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.280135 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xr6l\" (UniqueName: \"kubernetes.io/projected/58bc9d86-8ee5-4635-9de3-23890597f5ca-kube-api-access-8xr6l\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.280171 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.292726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.292796 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.292811 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.292833 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.292845 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.297495 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.311974 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.327366 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.343140 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.357420 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.370867 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.381697 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xr6l\" (UniqueName: \"kubernetes.io/projected/58bc9d86-8ee5-4635-9de3-23890597f5ca-kube-api-access-8xr6l\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.381752 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.381909 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.381984 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:44:14.881960406 +0000 UTC m=+36.509050674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.395953 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.395998 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.396009 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.396027 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.396038 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.400784 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xr6l\" (UniqueName: \"kubernetes.io/projected/58bc9d86-8ee5-4635-9de3-23890597f5ca-kube-api-access-8xr6l\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.404351 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" event={"ID":"18ae2562-beb3-4423-9045-f8559d492d8a","Type":"ContainerStarted","Data":"f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.404420 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" event={"ID":"18ae2562-beb3-4423-9045-f8559d492d8a","Type":"ContainerStarted","Data":"22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.421544 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.435181 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.448014 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.464600 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.478361 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.491474 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.499131 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.499202 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.499216 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.499237 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.499250 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.506121 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.521838 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.536631 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.551769 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.565514 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.586238 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.602359 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.602400 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.602417 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.602348 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.602440 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.602453 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.616349 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.630296 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.643953 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.663505 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.705702 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.705759 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.705771 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.705798 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.705815 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.808753 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.808805 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.808817 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.808836 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.808848 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.856208 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.856272 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.856285 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.856307 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.856322 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.870797 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.876500 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.876577 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.876597 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.876621 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.876660 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.887367 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.887571 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.887669 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:44:15.887628543 +0000 UTC m=+37.514718811 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.891336 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.895507 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.895568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.895582 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.895602 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.895616 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.908431 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.912288 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.912338 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.912352 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.912409 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.912426 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.925043 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.929171 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.929225 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.929248 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.929275 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.929290 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.943454 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:14Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:14 crc kubenswrapper[4675]: E0219 08:44:14.943630 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.946150 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.946217 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.946229 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.946251 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:14 crc kubenswrapper[4675]: I0219 08:44:14.946265 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:14Z","lastTransitionTime":"2026-02-19T08:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.049180 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.049244 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.049258 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.049361 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.049387 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.068778 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:02:46.084035237 +0000 UTC Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.102942 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:15 crc kubenswrapper[4675]: E0219 08:44:15.103210 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.152684 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.152746 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.152761 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.152786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.152802 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.256673 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.256740 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.256753 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.256775 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.256790 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.359391 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.359449 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.359461 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.359488 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.359501 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.462110 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.462207 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.462222 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.462244 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.462256 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.565319 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.565406 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.565418 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.565441 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.565455 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.669422 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.669511 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.669530 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.669571 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.669596 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.772907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.772969 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.772986 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.773009 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.773026 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.875930 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.875985 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.875996 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.876014 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.876029 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.898391 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:15 crc kubenswrapper[4675]: E0219 08:44:15.898606 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:15 crc kubenswrapper[4675]: E0219 08:44:15.898695 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:44:17.898677387 +0000 UTC m=+39.525767675 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.978580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.978676 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.978689 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.978712 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:15 crc kubenswrapper[4675]: I0219 08:44:15.978729 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:15Z","lastTransitionTime":"2026-02-19T08:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.069162 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 03:27:13.473835358 +0000 UTC Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.081791 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.081857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.081870 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.081914 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.081929 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.086440 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.103070 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.103068 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.103218 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:16 crc kubenswrapper[4675]: E0219 08:44:16.103345 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.103359 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:16 crc kubenswrapper[4675]: E0219 08:44:16.103472 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:16 crc kubenswrapper[4675]: E0219 08:44:16.103568 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.121186 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.135710 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.151705 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.170336 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.184768 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.184825 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.184843 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.184869 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.184889 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.187946 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.211139 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.224141 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.246291 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.259632 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.272411 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.284763 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.287340 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.287386 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.287400 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.287423 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.287438 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.297091 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.312723 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.328531 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.344711 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.360209 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:16Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.391081 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.391138 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.391154 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.391179 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.391196 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.493949 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.494012 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.494024 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.494044 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.494056 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.596839 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.596900 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.596917 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.596938 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.596953 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.699520 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.699568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.699579 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.699596 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.699606 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.803523 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.803603 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.803619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.803659 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.803674 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.907195 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.907273 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.907318 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.907353 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:16 crc kubenswrapper[4675]: I0219 08:44:16.907378 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:16Z","lastTransitionTime":"2026-02-19T08:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.010672 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.010734 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.010749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.010771 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.010785 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.069570 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 01:15:11.354509138 +0000 UTC Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.102985 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:17 crc kubenswrapper[4675]: E0219 08:44:17.103124 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.115160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.115226 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.115239 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.115264 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.115277 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.218251 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.218296 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.218305 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.218322 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.218334 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.321022 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.321075 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.321086 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.321104 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.321117 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.424162 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.424223 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.424243 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.424289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.424320 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.528035 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.528117 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.528140 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.528171 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.528198 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.631214 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.631295 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.631312 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.631428 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.631461 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.734769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.734821 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.734834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.734857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.734873 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.837873 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.837920 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.837930 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.837950 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.837962 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.925117 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:17 crc kubenswrapper[4675]: E0219 08:44:17.925252 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:17 crc kubenswrapper[4675]: E0219 08:44:17.925311 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:44:21.92529641 +0000 UTC m=+43.552386678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.940992 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.941044 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.941062 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.941083 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:17 crc kubenswrapper[4675]: I0219 08:44:17.941099 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:17Z","lastTransitionTime":"2026-02-19T08:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.044502 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.044567 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.044586 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.044612 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.044630 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.070066 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 07:28:45.391862435 +0000 UTC Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.102536 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.102629 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.102622 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:18 crc kubenswrapper[4675]: E0219 08:44:18.102848 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:18 crc kubenswrapper[4675]: E0219 08:44:18.103006 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:18 crc kubenswrapper[4675]: E0219 08:44:18.103162 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.147594 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.147719 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.147739 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.147764 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.147793 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.251457 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.251621 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.251687 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.251740 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.251766 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.353805 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.353895 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.353909 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.353925 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.353935 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.456139 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.456187 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.456197 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.456216 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.456231 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.558847 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.558899 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.558908 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.558933 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.558943 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.661552 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.661618 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.661667 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.661697 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.661715 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.764872 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.764941 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.764962 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.764988 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.765007 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.867232 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.867275 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.867284 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.867298 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.867308 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.970474 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.970576 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.970601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.970632 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:18 crc kubenswrapper[4675]: I0219 08:44:18.970701 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:18Z","lastTransitionTime":"2026-02-19T08:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.070904 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 07:22:37.590829422 +0000 UTC Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.073597 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.073692 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.073706 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.073723 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.073734 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.102752 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:19 crc kubenswrapper[4675]: E0219 08:44:19.102909 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.124200 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.144207 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.165603 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.176957 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.177003 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.177021 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.177046 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.177063 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.185194 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.202569 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.225045 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.242400 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.254669 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.278709 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.278748 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.278806 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.278823 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.278833 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.283014 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28df5155d34cf71a1035d54a841196d2f96c786207c71f619754a7f40389a23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"message\\\":\\\"s/factory.go:160\\\\nI0219 08:44:10.182126 5887 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0219 08:44:10.182169 5887 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:10.182108 5887 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0219 08:44:10.182415 5887 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0219 08:44:10.182519 5887 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0219 08:44:10.182863 5887 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0219 08:44:10.182906 5887 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:10.182937 5887 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0219 08:44:10.182951 5887 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:10.182959 5887 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0219 08:44:10.182975 5887 factory.go:656] Stopping watch factory\\\\nI0219 08:44:10.182979 5887 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:10.182992 5887 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:10.182996 5887 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0219 08:44:10.183018 5887 metrics.go:553] Stopping metrics server at address \\\\\\\"\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.294243 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.305941 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.321187 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.338239 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.353879 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.373534 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.380905 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.380964 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.380982 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.381008 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.381027 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.389068 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.402047 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.483588 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.483953 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.484121 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.484302 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.484499 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.588761 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.588823 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.588835 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.588855 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.588869 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.692022 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.692089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.692106 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.692134 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.692151 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.796910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.797350 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.797422 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.797494 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.797579 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.900579 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.900619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.900630 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.900670 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:19 crc kubenswrapper[4675]: I0219 08:44:19.900692 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:19Z","lastTransitionTime":"2026-02-19T08:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.003304 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.003378 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.003400 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.003429 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.003447 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.071318 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 19:34:22.4899169 +0000 UTC Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.103007 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.103056 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.103216 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:20 crc kubenswrapper[4675]: E0219 08:44:20.103204 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:20 crc kubenswrapper[4675]: E0219 08:44:20.103396 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:20 crc kubenswrapper[4675]: E0219 08:44:20.103477 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.106085 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.106139 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.106178 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.106203 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.106221 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.209117 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.209179 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.209194 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.209217 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.209231 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.313133 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.313181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.313211 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.313236 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.313248 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.415831 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.415879 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.415891 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.415910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.415921 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.519589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.519654 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.519665 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.519685 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.519698 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.623169 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.623221 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.623233 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.623266 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.623282 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.726345 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.726381 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.726393 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.726412 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.726424 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.828564 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.828622 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.828666 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.828689 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.828702 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.930562 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.930625 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.930663 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.930684 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:20 crc kubenswrapper[4675]: I0219 08:44:20.930697 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:20Z","lastTransitionTime":"2026-02-19T08:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.033803 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.033855 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.033865 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.033891 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.033905 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.072383 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 04:24:22.124494322 +0000 UTC Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.102962 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:21 crc kubenswrapper[4675]: E0219 08:44:21.103197 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.137208 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.137276 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.137289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.137313 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.137327 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.240498 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.240548 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.240563 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.240584 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.240597 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.344162 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.344223 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.344236 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.344257 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.344269 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.447595 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.447676 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.447687 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.447707 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.447717 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.551307 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.551363 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.551376 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.551400 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.551415 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.655292 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.655348 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.655362 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.655386 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.655402 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.759356 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.759416 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.759430 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.759450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.759460 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.862736 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.862816 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.862826 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.862845 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.862857 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.966185 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.966238 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.966250 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.966272 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.966286 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:21Z","lastTransitionTime":"2026-02-19T08:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:21 crc kubenswrapper[4675]: I0219 08:44:21.969547 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:21 crc kubenswrapper[4675]: E0219 08:44:21.971128 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:21 crc kubenswrapper[4675]: E0219 08:44:21.971287 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:44:29.971238969 +0000 UTC m=+51.598329237 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.068731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.068784 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.068796 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.068817 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.068830 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.072517 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 08:13:04.418655105 +0000 UTC Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.102936 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:22 crc kubenswrapper[4675]: E0219 08:44:22.103129 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.102981 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:22 crc kubenswrapper[4675]: E0219 08:44:22.103228 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.102954 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:22 crc kubenswrapper[4675]: E0219 08:44:22.103285 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.171739 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.171799 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.171813 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.171833 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.171849 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.276028 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.276096 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.276111 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.276142 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.276164 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.379238 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.379281 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.379289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.379308 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.379318 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.481779 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.481825 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.481856 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.481878 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.481889 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.584938 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.585000 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.585014 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.585036 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.585054 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.687067 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.687106 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.687117 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.687134 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.687145 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.790021 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.790089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.790118 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.790141 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.790157 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.893907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.893965 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.893979 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.894000 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.894015 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.996905 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.996960 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.996972 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.996992 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:22 crc kubenswrapper[4675]: I0219 08:44:22.997005 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:22Z","lastTransitionTime":"2026-02-19T08:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.073378 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 11:43:09.990543133 +0000 UTC Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.099691 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.099763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.099774 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.099791 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.099804 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.103129 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:23 crc kubenswrapper[4675]: E0219 08:44:23.103615 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.103924 4675 scope.go:117] "RemoveContainer" containerID="d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.119699 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.141666 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.157539 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.169806 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.183082 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.198743 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.203419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.203462 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.203472 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.203489 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.203500 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.214822 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.228778 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.244317 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.265585 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.284634 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.300372 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.307146 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.307189 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.307204 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.307227 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.307241 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.314778 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.334597 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.346269 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.359616 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.371673 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.410697 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.410741 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.410751 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.410767 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.410779 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.435715 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/1.log" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.438875 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.439024 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.455536 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.468554 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.485237 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.503711 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.513421 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.513476 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.513524 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.513547 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.513560 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.521760 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.546022 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.569690 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.590764 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.605031 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.616487 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.616515 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.616524 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.616541 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.616553 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.621026 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.636109 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.652390 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.671929 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.700423 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.719531 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.719575 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.719588 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.719611 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.719625 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.730418 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.757182 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.774755 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:23Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.822366 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.822432 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.822444 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.822467 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.822497 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.928167 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.928245 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.928263 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.928292 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:23 crc kubenswrapper[4675]: I0219 08:44:23.928314 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:23Z","lastTransitionTime":"2026-02-19T08:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.031929 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.031979 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.031989 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.032009 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.032020 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.074230 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 05:36:08.483588389 +0000 UTC Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.102985 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.103101 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.103175 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:24 crc kubenswrapper[4675]: E0219 08:44:24.103345 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:24 crc kubenswrapper[4675]: E0219 08:44:24.103494 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:24 crc kubenswrapper[4675]: E0219 08:44:24.103614 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.135580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.135629 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.135666 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.135694 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.135706 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.238694 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.238739 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.238751 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.238767 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.238783 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.341626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.341699 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.341712 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.341731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.341744 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.443877 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.443930 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.443940 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.443962 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.443975 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.444774 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/2.log" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.445695 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/1.log" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.449025 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4" exitCode=1 Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.449072 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.449112 4675 scope.go:117] "RemoveContainer" containerID="d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.450242 4675 scope.go:117] "RemoveContainer" containerID="fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4" Feb 19 08:44:24 crc kubenswrapper[4675]: E0219 08:44:24.450481 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.480518 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.495024 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.507612 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.522044 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.538031 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.546951 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.547001 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.547011 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.547028 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.547040 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.551807 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.566733 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.581878 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.601417 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.614312 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.625438 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.640597 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.650092 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.650151 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.650163 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.650182 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.650195 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.651962 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.673343 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d54230968a004d6f2c4bde74d4d2742866e007b3df69e369f087f37908c4b7b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:11Z\\\",\\\"message\\\":\\\"ion{},},}\\\\nI0219 08:44:11.403290 6070 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0219 08:44:11.403359 6070 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nF0219 08:44:11.403361 6070 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:11Z is after 2025-08-24T17:21:41Z]\\\\nI0219 08:44:11.403368 6070 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-lkvd2\\\\nI0219 08:44:11.403375 6070 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=defaul\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.684868 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.697220 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.709262 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:24Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.752974 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.753044 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.753090 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.753117 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.753132 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.855974 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.856045 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.856057 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.856075 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.856089 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.959099 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.959150 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.959159 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.959176 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:24 crc kubenswrapper[4675]: I0219 08:44:24.959187 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:24Z","lastTransitionTime":"2026-02-19T08:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.062086 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.062155 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.062176 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.062202 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.062213 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.074477 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 04:10:07.354937676 +0000 UTC Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.103259 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:25 crc kubenswrapper[4675]: E0219 08:44:25.103483 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.165785 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.166109 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.166192 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.166288 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.166373 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.269690 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.269744 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.269754 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.269770 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.269781 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.277437 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.277491 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.277504 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.277527 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.277542 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: E0219 08:44:25.294874 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:25Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.300491 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.300560 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.300588 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.300707 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.300722 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: E0219 08:44:25.315503 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:25Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.320062 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.320098 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.320109 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.320126 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.320137 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: E0219 08:44:25.338444 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:25Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.344316 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.344399 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.344425 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.344455 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.344479 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: E0219 08:44:25.360793 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:25Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.365544 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.365601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.365616 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.365667 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.365715 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: E0219 08:44:25.377718 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:25Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:25 crc kubenswrapper[4675]: E0219 08:44:25.377888 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.379699 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.379764 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.379786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.379819 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.379841 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.456632 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/2.log" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.483024 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.483080 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.483090 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.483135 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.483155 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.586119 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.586220 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.586233 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.586250 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.586291 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.689545 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.689614 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.689632 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.689696 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.689715 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.793074 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.793156 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.793182 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.793213 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.793234 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.896809 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.896891 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.896922 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.896950 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.896971 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.999186 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.999573 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.999672 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.999769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:25 crc kubenswrapper[4675]: I0219 08:44:25.999844 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:25Z","lastTransitionTime":"2026-02-19T08:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.075435 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 15:37:06.224094341 +0000 UTC Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.102475 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.102596 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:26 crc kubenswrapper[4675]: E0219 08:44:26.102863 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.103017 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.103062 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: E0219 08:44:26.103054 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.103078 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.103146 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.103172 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.103876 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:26 crc kubenswrapper[4675]: E0219 08:44:26.104212 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.206058 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.206415 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.206559 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.206762 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.206945 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.309588 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.309710 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.309737 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.309770 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.309795 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.412323 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.412359 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.412369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.412383 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.412392 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.514257 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.514303 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.514315 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.514337 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.514353 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.616368 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.616395 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.616402 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.616440 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.616449 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.719374 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.719427 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.719436 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.719453 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.719462 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.822386 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.822472 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.822486 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.822526 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.822579 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.925992 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.926066 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.926080 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.926112 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:26 crc kubenswrapper[4675]: I0219 08:44:26.926124 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:26Z","lastTransitionTime":"2026-02-19T08:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.028573 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.028670 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.028691 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.028715 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.028732 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.075853 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 23:55:03.226020285 +0000 UTC Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.103292 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:27 crc kubenswrapper[4675]: E0219 08:44:27.103444 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.131709 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.131758 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.131769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.131788 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.131801 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.234692 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.234736 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.234747 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.234766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.234780 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.338104 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.338155 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.338165 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.338192 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.338206 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.441667 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.441725 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.441739 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.441755 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.441764 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.545184 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.545234 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.545252 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.545271 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.545285 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.648682 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.648751 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.648763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.648786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.648804 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.751295 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.751340 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.751360 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.751382 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.751399 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.854401 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.854455 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.854471 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.854492 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.854505 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.957527 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.957588 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.957598 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.957619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:27 crc kubenswrapper[4675]: I0219 08:44:27.957633 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:27Z","lastTransitionTime":"2026-02-19T08:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.060776 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.060828 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.060839 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.060858 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.060871 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.076170 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 22:56:00.555838028 +0000 UTC Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.102567 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.102757 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:28 crc kubenswrapper[4675]: E0219 08:44:28.102882 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.102776 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:28 crc kubenswrapper[4675]: E0219 08:44:28.102991 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:28 crc kubenswrapper[4675]: E0219 08:44:28.103019 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.164096 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.164146 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.164160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.164188 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.164201 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.202999 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.204599 4675 scope.go:117] "RemoveContainer" containerID="fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4" Feb 19 08:44:28 crc kubenswrapper[4675]: E0219 08:44:28.204891 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.218759 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.232687 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.246797 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.266659 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.266746 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.266760 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.266780 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.266795 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.273195 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.289888 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.303604 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.317392 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.335866 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.351794 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.366115 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.369399 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.369424 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.369432 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.369446 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.369455 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.379285 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.400023 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.416458 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.430136 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.447007 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.463427 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.471070 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.471122 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.471133 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.471153 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.471165 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.493447 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:28Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.575283 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.575358 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.575392 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.575423 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.575491 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.678594 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.678690 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.678708 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.678733 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.678751 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.781408 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.781483 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.781514 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.781542 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.781565 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.884356 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.884427 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.884438 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.884458 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.884470 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.987450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.987496 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.987506 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.987525 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:28 crc kubenswrapper[4675]: I0219 08:44:28.987536 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:28Z","lastTransitionTime":"2026-02-19T08:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.077178 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 00:07:08.966186165 +0000 UTC Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.091001 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.091056 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.091067 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.091088 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.091102 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.102410 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:29 crc kubenswrapper[4675]: E0219 08:44:29.102594 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.124168 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.150387 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.164907 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.176728 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.189181 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.193491 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.193530 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.193542 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.193566 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.193603 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.206928 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.221919 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.237101 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.251542 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.271087 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.286532 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.296726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.296776 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.296796 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.296821 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.296839 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.303749 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.318501 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.322479 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.335111 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.340397 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.353987 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.371411 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.388264 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.404569 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.404628 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.404658 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.404682 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.404702 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.407073 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.428487 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.443813 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.459130 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.475676 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.496798 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.508394 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.508456 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.508466 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.508485 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.508497 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.512358 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.527317 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.540767 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.554606 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.570078 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.585361 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.601018 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.611056 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.611458 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.611527 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.611686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.611776 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.616522 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.638777 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.655945 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.671466 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.683695 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:29Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.715274 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.715340 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.715353 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.715369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.715381 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.818575 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.818617 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.818627 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.818665 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.818684 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.920724 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.920792 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.920807 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.920837 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.920853 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:29Z","lastTransitionTime":"2026-02-19T08:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:29 crc kubenswrapper[4675]: I0219 08:44:29.974994 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:29 crc kubenswrapper[4675]: E0219 08:44:29.975199 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:29 crc kubenswrapper[4675]: E0219 08:44:29.975334 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:44:45.975310741 +0000 UTC m=+67.602401029 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.023566 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.023708 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.023738 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.023767 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.023790 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.078096 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 16:38:59.505582824 +0000 UTC Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.102418 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.102499 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.102580 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.102847 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.102972 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.103085 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.125924 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.125963 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.125976 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.125993 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.126005 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.177098 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.177270 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177318 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:45:02.17728913 +0000 UTC m=+83.804379438 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.177360 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.177420 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177440 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177467 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177491 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.177513 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177562 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:45:02.177539207 +0000 UTC m=+83.804629515 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177616 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177715 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:45:02.177698181 +0000 UTC m=+83.804788489 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177793 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177832 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:45:02.177820074 +0000 UTC m=+83.804910382 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.177997 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.178037 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.178052 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:30 crc kubenswrapper[4675]: E0219 08:44:30.178123 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:45:02.178104132 +0000 UTC m=+83.805194410 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.229105 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.229165 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.229180 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.229201 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.229213 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.332372 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.332438 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.332449 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.332473 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.332487 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.435289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.435339 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.435357 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.435380 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.435398 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.538022 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.538085 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.538103 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.538123 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.538141 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.640034 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.640085 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.640096 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.640155 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.640170 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.742494 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.742543 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.742556 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.742576 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.742588 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.845227 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.845294 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.845310 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.845334 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.845348 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.948807 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.948857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.948880 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.948911 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:30 crc kubenswrapper[4675]: I0219 08:44:30.948934 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:30Z","lastTransitionTime":"2026-02-19T08:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.052113 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.052184 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.052195 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.052224 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.052239 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.078518 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 00:20:14.420282503 +0000 UTC Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.103325 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:31 crc kubenswrapper[4675]: E0219 08:44:31.103570 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.155328 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.155371 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.155382 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.155402 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.155416 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.258446 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.258519 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.258534 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.258562 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.258581 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.362118 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.362168 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.362218 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.362239 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.362277 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.465098 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.465153 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.465165 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.465185 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.465197 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.568635 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.568816 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.568887 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.568973 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.569045 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.671816 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.671873 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.671886 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.671907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.671921 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.775489 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.775961 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.776064 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.776146 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.776210 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.879839 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.879908 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.879925 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.879948 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.879962 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.983244 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.983303 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.983322 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.983343 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:31 crc kubenswrapper[4675]: I0219 08:44:31.983358 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:31Z","lastTransitionTime":"2026-02-19T08:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.079103 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 15:57:09.984823941 +0000 UTC Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.085810 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.085874 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.085884 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.085898 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.085908 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.103120 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.103207 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.103304 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:32 crc kubenswrapper[4675]: E0219 08:44:32.103412 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:32 crc kubenswrapper[4675]: E0219 08:44:32.103519 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:32 crc kubenswrapper[4675]: E0219 08:44:32.103663 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.188606 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.188683 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.188698 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.188724 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.188738 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.292539 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.292602 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.292617 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.292656 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.292673 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.396449 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.396508 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.396518 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.396537 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.396549 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.499093 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.499152 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.499164 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.499181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.499192 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.601461 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.601506 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.601517 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.601536 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.601547 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.704589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.704686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.704703 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.704731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.704746 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.807493 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.807548 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.807560 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.807581 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.807594 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.910831 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.910898 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.910909 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.910933 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:32 crc kubenswrapper[4675]: I0219 08:44:32.910949 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:32Z","lastTransitionTime":"2026-02-19T08:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.013284 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.013336 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.013347 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.013363 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.013375 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.079325 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 19:08:58.228202846 +0000 UTC Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.103139 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:33 crc kubenswrapper[4675]: E0219 08:44:33.103319 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.116019 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.116079 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.116099 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.116127 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.116142 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.218896 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.218930 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.218941 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.218958 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.218969 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.322101 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.322150 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.322159 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.322177 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.322189 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.424941 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.425003 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.425014 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.425030 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.425042 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.527500 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.527538 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.527547 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.527563 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.527574 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.635474 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.636383 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.636428 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.636458 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.636479 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.739367 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.739475 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.739493 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.739513 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.739525 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.842223 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.842683 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.842848 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.843019 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.843150 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.945929 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.945969 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.945979 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.945993 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:33 crc kubenswrapper[4675]: I0219 08:44:33.946002 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:33Z","lastTransitionTime":"2026-02-19T08:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.049324 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.049390 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.049415 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.049447 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.049471 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.080068 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 01:37:42.0823445 +0000 UTC Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.102799 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.102813 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:34 crc kubenswrapper[4675]: E0219 08:44:34.103076 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.102844 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:34 crc kubenswrapper[4675]: E0219 08:44:34.103235 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:34 crc kubenswrapper[4675]: E0219 08:44:34.103434 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.152030 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.152095 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.152110 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.152132 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.152150 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.255385 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.255437 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.255450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.255469 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.255480 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.358356 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.358403 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.358414 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.358435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.358448 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.460650 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.460697 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.460711 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.460730 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.460744 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.563631 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.563733 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.563745 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.563769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.563786 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.666468 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.666524 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.666533 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.666549 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.666559 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.769514 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.769565 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.769574 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.769590 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.769601 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.873252 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.873375 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.873395 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.873421 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.873511 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.977441 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.977484 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.977492 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.977516 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:34 crc kubenswrapper[4675]: I0219 08:44:34.977536 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:34Z","lastTransitionTime":"2026-02-19T08:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.079778 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.079825 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.079834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.079850 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.079860 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.080534 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 05:36:01.912137012 +0000 UTC Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.102281 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:35 crc kubenswrapper[4675]: E0219 08:44:35.102417 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.183285 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.183344 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.183382 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.183410 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.183427 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.286943 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.287018 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.287027 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.287051 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.287066 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.390562 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.390603 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.390615 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.390649 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.390661 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.494261 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.494293 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.494304 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.494319 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.494328 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.597112 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.597235 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.597246 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.597267 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.597279 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.700520 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.700568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.700579 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.700601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.700612 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.754166 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.754220 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.754231 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.754251 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.754263 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: E0219 08:44:35.770660 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:35Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.774809 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.774858 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.774867 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.774882 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.774891 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: E0219 08:44:35.789709 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:35Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.796517 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.796570 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.796588 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.796610 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.796624 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: E0219 08:44:35.814934 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:35Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.818442 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.818499 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.818511 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.818534 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.818548 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: E0219 08:44:35.831001 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:35Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.835340 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.835398 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.835416 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.835441 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.835456 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: E0219 08:44:35.848356 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:35Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:35 crc kubenswrapper[4675]: E0219 08:44:35.848518 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.850358 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.850408 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.850419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.850438 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.850450 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.954295 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.954336 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.954344 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.954367 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:35 crc kubenswrapper[4675]: I0219 08:44:35.954379 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:35Z","lastTransitionTime":"2026-02-19T08:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.057174 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.057217 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.057225 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.057237 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.057246 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.080714 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 03:53:05.376871885 +0000 UTC Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.103344 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.103375 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:36 crc kubenswrapper[4675]: E0219 08:44:36.103551 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:36 crc kubenswrapper[4675]: E0219 08:44:36.103759 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.103378 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:36 crc kubenswrapper[4675]: E0219 08:44:36.103898 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.160212 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.160272 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.160289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.160307 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.160320 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.263713 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.263769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.263783 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.263804 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.263819 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.367209 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.367324 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.367341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.367369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.367387 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.470735 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.470783 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.470795 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.470817 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.470828 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.574144 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.574193 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.574203 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.574222 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.574234 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.676831 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.676937 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.676970 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.677008 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.677029 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.780630 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.780731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.780786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.780817 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.780836 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.884052 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.884108 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.884122 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.884141 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.884153 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.987696 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.987808 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.987825 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.987843 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:36 crc kubenswrapper[4675]: I0219 08:44:36.987856 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:36Z","lastTransitionTime":"2026-02-19T08:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.081334 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 23:59:06.45711542 +0000 UTC Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.090866 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.090921 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.090938 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.090961 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.090975 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.102400 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:37 crc kubenswrapper[4675]: E0219 08:44:37.102545 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.194315 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.194378 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.194399 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.194430 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.194448 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.297206 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.297260 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.297273 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.297294 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.297305 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.400391 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.400434 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.400444 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.400462 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.400473 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.502668 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.502754 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.502771 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.502797 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.502812 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.606435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.606479 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.606492 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.606510 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.606522 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.711177 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.711237 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.711249 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.711270 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.711283 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.814227 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.814274 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.814282 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.814301 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.814313 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.917815 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.917860 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.917871 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.917890 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:37 crc kubenswrapper[4675]: I0219 08:44:37.917900 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:37Z","lastTransitionTime":"2026-02-19T08:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.021160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.021205 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.021217 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.021237 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.021250 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.082255 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 11:42:37.969825608 +0000 UTC Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.102841 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:38 crc kubenswrapper[4675]: E0219 08:44:38.103026 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.103269 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:38 crc kubenswrapper[4675]: E0219 08:44:38.103342 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.103513 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:38 crc kubenswrapper[4675]: E0219 08:44:38.103584 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.124898 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.124948 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.124961 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.124983 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.124997 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.228344 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.228394 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.228404 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.228423 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.228436 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.331526 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.331574 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.331584 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.331605 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.331615 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.434913 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.434983 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.434993 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.435013 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.435027 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.538117 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.538167 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.538177 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.538198 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.538211 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.641202 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.641254 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.641266 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.641285 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.641297 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.743897 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.743947 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.743959 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.743976 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.743986 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.846519 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.846567 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.846579 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.846594 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.846604 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.949237 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.949286 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.949297 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.949316 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:38 crc kubenswrapper[4675]: I0219 08:44:38.949328 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:38Z","lastTransitionTime":"2026-02-19T08:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.052063 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.052208 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.052229 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.052295 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.052314 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.083458 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 10:37:31.373378355 +0000 UTC Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.102621 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:39 crc kubenswrapper[4675]: E0219 08:44:39.103293 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.103688 4675 scope.go:117] "RemoveContainer" containerID="fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4" Feb 19 08:44:39 crc kubenswrapper[4675]: E0219 08:44:39.104086 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.120296 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.135018 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.148029 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.156603 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.156664 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.156675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.156699 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.156715 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.165876 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.182630 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.203060 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.217984 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.230228 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.250143 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.260570 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.260675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.260721 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.260890 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.260919 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.262911 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.279496 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.293992 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.309247 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.324706 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.341600 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.356081 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.364086 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.364137 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.364150 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.364171 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.364184 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.370302 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.382475 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:39Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.467548 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.467601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.467611 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.467626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.467657 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.570552 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.570626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.570927 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.571288 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.571341 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.674295 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.674363 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.674389 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.674419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.674443 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.778154 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.778214 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.778225 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.778263 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.778276 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.881599 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.881669 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.881684 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.881703 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.881715 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.984834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.984909 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.984939 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.984960 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:39 crc kubenswrapper[4675]: I0219 08:44:39.984973 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:39Z","lastTransitionTime":"2026-02-19T08:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.084410 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 23:05:48.500920122 +0000 UTC Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.087808 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.087884 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.087898 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.087918 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.087930 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.102314 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:40 crc kubenswrapper[4675]: E0219 08:44:40.102441 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.102315 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.102547 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:40 crc kubenswrapper[4675]: E0219 08:44:40.102770 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:40 crc kubenswrapper[4675]: E0219 08:44:40.102961 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.190786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.190849 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.190862 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.190888 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.190905 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.293925 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.293965 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.293976 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.293993 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.294004 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.396761 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.396814 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.396825 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.396843 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.396856 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.499720 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.499771 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.499797 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.499822 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.499838 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.602617 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.602686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.602700 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.602718 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.602730 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.704810 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.704851 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.704860 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.704874 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.704883 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.807039 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.807082 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.807091 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.807107 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.807117 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.911319 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.911391 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.911402 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.911419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:40 crc kubenswrapper[4675]: I0219 08:44:40.911435 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:40Z","lastTransitionTime":"2026-02-19T08:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.014409 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.014460 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.014478 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.014500 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.014517 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.085594 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 09:02:15.302595276 +0000 UTC Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.103248 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:41 crc kubenswrapper[4675]: E0219 08:44:41.103443 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.116668 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.116721 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.116732 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.116749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.116760 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.220089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.220156 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.220167 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.220189 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.220203 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.323759 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.323789 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.323802 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.323817 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.323829 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.427038 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.427134 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.427152 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.427176 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.427192 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.529938 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.530246 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.530320 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.530404 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.530483 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.633611 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.633945 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.634090 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.634219 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.634318 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.737957 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.738018 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.738037 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.738062 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.738084 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.853217 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.853298 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.853317 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.853809 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.853846 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.956774 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.956830 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.956846 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.956870 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:41 crc kubenswrapper[4675]: I0219 08:44:41.956890 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:41Z","lastTransitionTime":"2026-02-19T08:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.059731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.060022 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.060125 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.060228 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.060316 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.086201 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 03:10:35.593656972 +0000 UTC Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.102547 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.102614 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.102680 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:42 crc kubenswrapper[4675]: E0219 08:44:42.103094 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:42 crc kubenswrapper[4675]: E0219 08:44:42.102993 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:42 crc kubenswrapper[4675]: E0219 08:44:42.103205 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.163385 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.163413 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.163421 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.163433 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.163442 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.265212 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.265258 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.265269 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.265284 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.265292 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.367802 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.367859 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.367871 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.367888 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.367903 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.470163 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.470199 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.470211 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.470226 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.470237 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.573099 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.573229 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.573271 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.573314 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.573327 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.675597 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.675695 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.675709 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.675734 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.675745 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.778269 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.778338 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.778354 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.778378 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.778396 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.881917 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.882187 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.882196 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.882222 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.882233 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.984464 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.984525 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.984539 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.984556 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:42 crc kubenswrapper[4675]: I0219 08:44:42.984569 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:42Z","lastTransitionTime":"2026-02-19T08:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.086507 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 16:41:13.483513836 +0000 UTC Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.087057 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.087113 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.087123 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.087141 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.087154 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.102687 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:43 crc kubenswrapper[4675]: E0219 08:44:43.102942 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.191397 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.191474 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.191489 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.191516 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.191534 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.294736 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.294785 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.294798 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.294818 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.294829 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.396960 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.397001 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.397013 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.397030 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.397039 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.500353 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.500401 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.500411 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.500427 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.500438 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.602187 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.602243 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.602257 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.602305 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.602316 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.706260 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.706332 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.706348 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.706375 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.706393 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.809681 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.810108 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.810289 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.810440 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.810569 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.913750 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.913794 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.913806 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.913824 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:43 crc kubenswrapper[4675]: I0219 08:44:43.913835 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:43Z","lastTransitionTime":"2026-02-19T08:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.016350 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.016399 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.016411 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.016429 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.016440 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.087469 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 10:50:44.857330403 +0000 UTC Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.102813 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.102922 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:44 crc kubenswrapper[4675]: E0219 08:44:44.103011 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:44 crc kubenswrapper[4675]: E0219 08:44:44.103183 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.103316 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:44 crc kubenswrapper[4675]: E0219 08:44:44.103510 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.119134 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.119471 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.119593 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.119709 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.119780 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.222882 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.223212 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.223312 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.223401 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.223474 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.325421 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.325460 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.325471 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.325486 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.325496 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.428359 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.428426 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.428441 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.428462 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.428474 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.531271 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.531338 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.531352 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.531375 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.531392 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.633777 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.634314 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.634450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.634539 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.634634 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.737770 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.737829 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.737841 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.737883 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.737894 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.841275 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.841341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.841356 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.841378 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.841870 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.944970 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.945014 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.945024 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.945042 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:44 crc kubenswrapper[4675]: I0219 08:44:44.945055 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:44Z","lastTransitionTime":"2026-02-19T08:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.047853 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.047890 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.047899 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.047914 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.047924 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.089398 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 18:24:34.63885504 +0000 UTC Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.102948 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:45 crc kubenswrapper[4675]: E0219 08:44:45.103121 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.150501 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.150560 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.150574 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.150593 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.151339 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.254560 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.254591 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.254600 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.254614 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.254626 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.356745 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.356823 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.356833 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.356855 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.356869 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.459838 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.459901 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.459911 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.459934 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.459944 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.562551 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.562620 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.562650 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.562674 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.562689 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.664702 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.664736 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.664749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.664766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.664775 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.767077 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.767110 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.767121 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.767136 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.767148 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.870108 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.870144 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.870151 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.870167 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.870176 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.973149 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.973199 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.973211 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.973227 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:45 crc kubenswrapper[4675]: I0219 08:44:45.973238 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:45Z","lastTransitionTime":"2026-02-19T08:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.061694 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.061870 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.062522 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:45:18.062496683 +0000 UTC m=+99.689587131 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.076201 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.076248 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.076258 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.076278 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.076292 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.090583 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 11:01:58.546220265 +0000 UTC Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.103019 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.103150 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.103251 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.103163 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.103388 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.103708 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.178601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.178703 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.178719 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.178741 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.178755 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.213212 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.213265 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.213277 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.213299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.213332 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.231538 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:46Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.236426 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.236596 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.236704 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.236796 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.236881 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.250707 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:46Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.255351 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.255455 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.255523 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.255610 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.255726 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.270766 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:46Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.275802 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.275850 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.275862 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.275883 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.275899 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.292892 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:46Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.297760 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.297857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.297875 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.297926 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.297948 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.313992 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:46Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:46 crc kubenswrapper[4675]: E0219 08:44:46.314158 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.315853 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.315878 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.315889 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.315907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.315917 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.418497 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.418534 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.418543 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.418558 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.418568 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.532615 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.532665 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.532675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.532692 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.532703 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.635940 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.635983 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.635994 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.636012 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.636025 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.739091 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.739127 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.739135 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.739154 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.739165 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.841791 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.841832 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.841877 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.841899 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.841909 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.944199 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.944232 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.944240 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.944254 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:46 crc kubenswrapper[4675]: I0219 08:44:46.944263 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:46Z","lastTransitionTime":"2026-02-19T08:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.047522 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.047559 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.047568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.047586 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.047597 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.090893 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 02:03:13.716258875 +0000 UTC Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.103928 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:47 crc kubenswrapper[4675]: E0219 08:44:47.104116 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.150693 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.150743 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.150755 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.150776 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.150789 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.253703 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.253744 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.253752 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.253770 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.253781 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.356367 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.356413 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.356424 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.356442 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.356454 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.459074 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.459131 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.459144 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.459165 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.459184 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.562033 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.562079 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.562091 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.562108 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.562120 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.664269 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.664305 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.664317 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.664335 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.664347 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.767434 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.767496 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.767510 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.767532 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.767550 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.870488 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.870549 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.870567 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.870590 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.870608 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.974360 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.974500 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.974522 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.974550 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:47 crc kubenswrapper[4675]: I0219 08:44:47.974568 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:47Z","lastTransitionTime":"2026-02-19T08:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.077854 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.077898 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.077910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.077933 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.077946 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.091502 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 12:16:45.989433432 +0000 UTC Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.102967 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.103066 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.103130 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:48 crc kubenswrapper[4675]: E0219 08:44:48.103126 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:48 crc kubenswrapper[4675]: E0219 08:44:48.103244 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:48 crc kubenswrapper[4675]: E0219 08:44:48.103343 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.180057 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.180099 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.180112 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.180131 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.180146 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.283663 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.283713 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.283729 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.283754 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.283771 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.386215 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.386258 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.386268 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.386285 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.386296 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.489257 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.489308 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.489316 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.489331 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.489341 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.592091 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.592147 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.592158 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.592180 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.592192 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.695543 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.695604 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.695615 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.695654 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.695668 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.799567 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.799632 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.799665 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.799711 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.799733 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.902321 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.902362 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.902373 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.902389 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:48 crc kubenswrapper[4675]: I0219 08:44:48.902402 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:48Z","lastTransitionTime":"2026-02-19T08:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.005555 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.005607 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.005620 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.005663 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.005677 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.091696 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 12:44:58.199252066 +0000 UTC Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.103156 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:49 crc kubenswrapper[4675]: E0219 08:44:49.103337 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.108598 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.108660 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.108669 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.108687 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.108699 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.127164 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.143283 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.159297 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.171336 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.185537 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.200735 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.211485 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.211511 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.211520 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.211536 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.211547 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.217167 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.230167 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.253579 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.266811 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.280569 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.294757 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.308167 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.314050 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.314095 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.314106 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.314147 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.314162 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.323380 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.336687 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.349785 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.362418 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.377869 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.417034 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.417085 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.417094 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.417108 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.417117 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.519254 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.519749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.519852 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.519948 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.520051 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.623318 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.623373 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.623393 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.623415 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.623428 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.726397 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.726446 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.726457 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.726474 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.726487 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.766521 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/0.log" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.766583 4675 generic.go:334] "Generic (PLEG): container finished" podID="4cc07c5a-23e4-478f-88a7-73d41e17a320" containerID="b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b" exitCode=1 Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.766620 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerDied","Data":"b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.767124 4675 scope.go:117] "RemoveContainer" containerID="b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.786657 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.804836 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.821384 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.829258 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.829885 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.830185 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.830340 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.830482 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.839686 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.854907 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.871310 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.895759 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.913138 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.933513 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.933562 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.933572 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.933592 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.933609 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:49Z","lastTransitionTime":"2026-02-19T08:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.936040 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.947432 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.959211 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.972153 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:49 crc kubenswrapper[4675]: I0219 08:44:49.986113 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.000309 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:49Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.015205 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.030807 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.036049 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.036105 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.036124 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.036147 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.036163 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.046802 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.060376 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.092782 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 03:12:46.366131965 +0000 UTC Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.103230 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:50 crc kubenswrapper[4675]: E0219 08:44:50.103380 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.103230 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:50 crc kubenswrapper[4675]: E0219 08:44:50.103476 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.103230 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:50 crc kubenswrapper[4675]: E0219 08:44:50.103550 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.139673 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.139716 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.139726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.139741 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.139752 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.242693 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.242759 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.242776 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.242801 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.242821 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.346412 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.346728 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.346822 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.346900 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.346969 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.449866 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.449916 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.449940 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.449963 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.449975 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.551824 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.551860 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.551872 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.551890 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.551901 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.654984 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.655031 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.655042 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.655062 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.655074 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.758237 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.758315 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.758328 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.758346 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.758357 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.772908 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/0.log" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.773000 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerStarted","Data":"413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.789878 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.813281 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.829380 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.841946 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.857099 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.861140 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.861172 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.861188 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.861213 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.861231 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.874923 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.902566 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.918079 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.936567 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.950100 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.963436 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.963465 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.963473 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.963489 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.963498 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:50Z","lastTransitionTime":"2026-02-19T08:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.963820 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.977199 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:50 crc kubenswrapper[4675]: I0219 08:44:50.992917 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:50Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.007329 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:51Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.019275 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:51Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.042248 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:51Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.059469 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:51Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.065728 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.065758 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.065769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.065788 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.065800 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.073674 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:51Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.093913 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 20:50:33.947035213 +0000 UTC Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.103367 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:51 crc kubenswrapper[4675]: E0219 08:44:51.103593 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.168435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.168830 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.168910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.169005 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.169082 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.273409 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.273446 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.273460 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.273491 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.273505 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.376605 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.376677 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.376688 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.376712 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.376724 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.479486 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.479518 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.479529 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.479547 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.479559 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.582341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.582394 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.582406 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.582426 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.582442 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.686306 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.686362 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.686375 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.686399 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.686411 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.789503 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.789556 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.789567 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.789590 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.789611 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.893098 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.893152 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.893162 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.893178 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.893188 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.996419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.996456 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.996466 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.996480 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:51 crc kubenswrapper[4675]: I0219 08:44:51.996490 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:51Z","lastTransitionTime":"2026-02-19T08:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.094626 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 10:27:58.101398784 +0000 UTC Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.098619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.098666 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.098676 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.098694 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.098708 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.102406 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:52 crc kubenswrapper[4675]: E0219 08:44:52.102571 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.102743 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:52 crc kubenswrapper[4675]: E0219 08:44:52.102800 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.102955 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:52 crc kubenswrapper[4675]: E0219 08:44:52.103168 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.104078 4675 scope.go:117] "RemoveContainer" containerID="fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.201897 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.201959 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.201976 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.202005 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.202024 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.305089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.305129 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.305140 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.305159 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.305172 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.408132 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.408170 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.408179 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.408197 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.408208 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.510416 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.510449 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.510459 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.510473 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.510483 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.612834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.612878 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.612887 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.612905 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.612923 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.715514 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.715553 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.715564 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.715578 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.715588 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.783813 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/2.log" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.785869 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.787154 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.808411 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.820276 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.820319 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.820329 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.820348 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.820359 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.827481 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.841533 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.856015 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.877440 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.891495 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.910282 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.923606 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.923675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.923829 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.923853 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.924129 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:52Z","lastTransitionTime":"2026-02-19T08:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.924766 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:52 crc kubenswrapper[4675]: I0219 08:44:52.980289 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:52Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.005463 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.020010 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.026479 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.026538 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.026555 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.026580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.026595 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.037985 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.057792 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.076075 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.094812 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 11:23:46.922006028 +0000 UTC Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.094976 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.102339 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:53 crc kubenswrapper[4675]: E0219 08:44:53.102496 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.108131 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.119772 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.130019 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.130055 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.130066 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.130104 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.130124 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.131563 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.233122 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.233161 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.233171 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.233189 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.233201 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.335505 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.335556 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.335568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.335598 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.335614 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.439078 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.439126 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.439138 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.439157 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.439170 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.542586 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.542692 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.542706 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.542728 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.542760 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.645363 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.645435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.645452 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.645474 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.645485 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.749034 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.749097 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.749111 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.749133 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.749147 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.793242 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/3.log" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.794174 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/2.log" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.797679 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d" exitCode=1 Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.797748 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.797809 4675 scope.go:117] "RemoveContainer" containerID="fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.798548 4675 scope.go:117] "RemoveContainer" containerID="1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d" Feb 19 08:44:53 crc kubenswrapper[4675]: E0219 08:44:53.798939 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.829579 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.847320 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.852226 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.852284 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.852296 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.852317 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.852329 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.862724 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.877734 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.894159 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.907982 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.923534 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.947430 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe5e097d7747af60186b800df317a8c3bf9bd680b75bf2a78b03473417f1cbe4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:24Z\\\",\\\"message\\\":\\\"te:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0219 08:44:24.315041 6278 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}\\\\nF0219 08:44:24.315083 6278 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:53Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0219 08:44:53.059973 6685 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:53.060252 6685 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0219 08:44:53.060742 6685 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:53.060771 6685 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:53.060815 6685 factory.go:656] Stopping watch factory\\\\nI0219 08:44:53.060815 6685 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:53.060835 6685 handler.go:208] Removed *v1.Node event handler 2\\\\nI0219 08:44:53.062927 6685 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0219 08:44:53.062952 6685 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0219 08:44:53.063021 6685 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:53.063059 6685 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0219 08:44:53.063170 6685 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.954378 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.954404 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.954414 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.954430 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.954440 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:53Z","lastTransitionTime":"2026-02-19T08:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.960772 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.973244 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.985944 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:53 crc kubenswrapper[4675]: I0219 08:44:53.996618 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:53Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.009136 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.021163 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.034103 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.048516 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.057128 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.057166 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.057176 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.057195 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.057206 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.062580 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.076660 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.095268 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 10:26:31.966188146 +0000 UTC Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.102678 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:54 crc kubenswrapper[4675]: E0219 08:44:54.102818 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.102898 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:54 crc kubenswrapper[4675]: E0219 08:44:54.102945 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.102997 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:54 crc kubenswrapper[4675]: E0219 08:44:54.103050 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.159901 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.159966 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.159987 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.160011 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.160029 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.263915 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.264007 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.264027 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.264058 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.264079 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.367834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.367920 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.367952 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.367981 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.368003 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.471225 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.471304 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.471316 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.471339 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.471354 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.575510 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.575602 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.575631 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.575686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.575724 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.679091 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.679277 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.679301 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.679333 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.679350 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.782065 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.782114 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.782123 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.782140 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.782152 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.804967 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/3.log" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.810563 4675 scope.go:117] "RemoveContainer" containerID="1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d" Feb 19 08:44:54 crc kubenswrapper[4675]: E0219 08:44:54.810795 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.830169 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.853313 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.871574 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.885585 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.885657 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.885669 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.885690 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.885703 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.888186 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.903452 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.918086 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.934418 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.950679 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.972239 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:53Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0219 08:44:53.059973 6685 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:53.060252 6685 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0219 08:44:53.060742 6685 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:53.060771 6685 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:53.060815 6685 factory.go:656] Stopping watch factory\\\\nI0219 08:44:53.060815 6685 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:53.060835 6685 handler.go:208] Removed *v1.Node event handler 2\\\\nI0219 08:44:53.062927 6685 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0219 08:44:53.062952 6685 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0219 08:44:53.063021 6685 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:53.063059 6685 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0219 08:44:53.063170 6685 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.986961 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.990249 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.990299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.990345 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.990369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:54 crc kubenswrapper[4675]: I0219 08:44:54.990382 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:54Z","lastTransitionTime":"2026-02-19T08:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:54.999934 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:54Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.012439 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:55Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.025792 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:55Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.038245 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:55Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.048089 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:55Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.059900 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:55Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.078536 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:55Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.093249 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.093284 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.093294 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.093310 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.093322 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.096233 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 17:25:25.556163179 +0000 UTC Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.100533 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:55Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.102448 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:55 crc kubenswrapper[4675]: E0219 08:44:55.102617 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.196180 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.196243 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.196253 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.196274 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.196289 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.299999 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.300075 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.300095 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.300125 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.300141 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.404049 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.404115 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.404132 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.404156 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.404174 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.508249 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.508367 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.508388 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.508413 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.508442 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.612153 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.612223 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.612241 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.612269 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.612285 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.715307 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.715355 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.715365 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.715386 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.715397 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.819132 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.819190 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.819202 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.819226 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.819242 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.922238 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.922299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.922312 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.922337 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:55 crc kubenswrapper[4675]: I0219 08:44:55.922353 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:55Z","lastTransitionTime":"2026-02-19T08:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.026174 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.026263 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.026292 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.026314 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.026326 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.097186 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 08:24:09.096584228 +0000 UTC Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.102919 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.102960 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.102942 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.103135 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.103236 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.103333 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.129050 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.129130 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.129143 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.129166 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.129177 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.232223 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.232308 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.232323 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.232348 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.232362 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.334944 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.334990 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.334999 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.335018 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.335031 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.437799 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.437846 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.437856 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.437877 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.437891 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.540371 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.540462 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.540481 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.540514 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.540529 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.542349 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.542413 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.542425 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.542447 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.542460 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.556420 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:56Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.560712 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.560756 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.560769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.560798 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.560809 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.574923 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:56Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.579116 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.579150 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.579179 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.579205 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.579223 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.592011 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:56Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.595741 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.595795 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.595806 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.595826 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.595839 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.608227 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:56Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.612539 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.612568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.612578 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.612597 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.612611 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.624748 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:56Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:56 crc kubenswrapper[4675]: E0219 08:44:56.624906 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.643114 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.643146 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.643161 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.643177 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.643187 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.748662 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.748728 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.748742 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.748759 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.748771 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.851545 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.851602 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.851611 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.851631 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.851676 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.954255 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.954310 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.954323 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.954345 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:56 crc kubenswrapper[4675]: I0219 08:44:56.954365 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:56Z","lastTransitionTime":"2026-02-19T08:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.057575 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.057635 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.057691 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.057713 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.057727 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.098035 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 12:02:12.50789275 +0000 UTC Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.102562 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:57 crc kubenswrapper[4675]: E0219 08:44:57.102862 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.160592 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.160659 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.160673 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.160690 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.160705 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.264209 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.264275 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.264286 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.264310 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.264321 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.367114 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.367177 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.367197 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.367220 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.367409 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.470433 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.470506 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.470519 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.470540 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.470554 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.574694 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.574736 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.574748 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.574766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.574780 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.677649 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.677691 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.677701 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.677718 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.677728 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.780153 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.780224 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.780242 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.780266 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.780282 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.882862 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.882910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.882919 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.882935 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.882946 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.985526 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.985592 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.985607 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.985629 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:57 crc kubenswrapper[4675]: I0219 08:44:57.985662 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:57Z","lastTransitionTime":"2026-02-19T08:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.088001 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.088044 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.088053 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.088073 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.088085 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.098336 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 05:38:28.892701007 +0000 UTC Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.102774 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.102806 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.102823 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:44:58 crc kubenswrapper[4675]: E0219 08:44:58.102943 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:44:58 crc kubenswrapper[4675]: E0219 08:44:58.103072 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:44:58 crc kubenswrapper[4675]: E0219 08:44:58.103142 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.190015 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.190060 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.190070 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.190082 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.190094 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.293164 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.293214 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.293233 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.293256 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.293273 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.396453 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.396518 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.396539 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.396567 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.396589 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.499575 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.499630 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.499685 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.499708 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.499725 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.603137 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.603220 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.603259 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.603299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.603335 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.706601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.706704 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.706731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.706760 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.706785 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.809904 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.809939 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.809950 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.810012 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.810046 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.912103 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.912155 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.912166 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.912185 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:58 crc kubenswrapper[4675]: I0219 08:44:58.912198 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:58Z","lastTransitionTime":"2026-02-19T08:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.014756 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.014860 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.014870 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.014892 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.014903 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.099187 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 09:03:57.691993175 +0000 UTC Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.102626 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:44:59 crc kubenswrapper[4675]: E0219 08:44:59.102792 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.118630 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.118723 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.118740 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.118766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.118782 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.124014 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.141298 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.159399 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.174131 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.189880 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.209747 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.222402 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.222435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.222465 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.222483 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.222495 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.225621 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.239724 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.251771 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.265258 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.280931 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.302517 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.319141 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.324508 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.324546 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.324559 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.324613 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.324629 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.335076 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.361670 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:53Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0219 08:44:53.059973 6685 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:53.060252 6685 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0219 08:44:53.060742 6685 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:53.060771 6685 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:53.060815 6685 factory.go:656] Stopping watch factory\\\\nI0219 08:44:53.060815 6685 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:53.060835 6685 handler.go:208] Removed *v1.Node event handler 2\\\\nI0219 08:44:53.062927 6685 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0219 08:44:53.062952 6685 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0219 08:44:53.063021 6685 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:53.063059 6685 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0219 08:44:53.063170 6685 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.375444 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.390695 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.404097 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:44:59Z is after 2025-08-24T17:21:41Z" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.428367 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.428419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.428430 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.428450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.428463 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.531529 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.532189 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.532203 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.532224 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.532237 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.634681 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.634745 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.634764 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.634790 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.634810 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.737806 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.738308 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.738390 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.738627 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.738788 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.842547 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.842600 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.842610 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.842678 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.842693 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.945545 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.945592 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.945605 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.945621 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:44:59 crc kubenswrapper[4675]: I0219 08:44:59.945634 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:44:59Z","lastTransitionTime":"2026-02-19T08:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.048586 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.048632 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.048664 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.048680 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.048692 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.100101 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 06:10:29.134460966 +0000 UTC Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.102514 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:00 crc kubenswrapper[4675]: E0219 08:45:00.102726 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.103034 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.103323 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:00 crc kubenswrapper[4675]: E0219 08:45:00.104276 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:00 crc kubenswrapper[4675]: E0219 08:45:00.104414 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.152511 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.152604 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.152619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.152689 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.152708 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.255880 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.255918 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.255928 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.255942 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.255951 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.358154 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.358200 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.358209 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.358225 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.358239 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.460804 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.460883 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.460895 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.460919 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.460934 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.565214 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.565271 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.565286 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.565313 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.565326 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.668444 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.668496 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.668507 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.668528 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.668540 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.772326 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.772382 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.772394 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.772416 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.772430 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.877029 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.877069 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.877078 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.877098 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.877108 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.979617 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.979681 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.979695 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.979717 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:00 crc kubenswrapper[4675]: I0219 08:45:00.979730 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:00Z","lastTransitionTime":"2026-02-19T08:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.082187 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.082245 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.082261 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.082285 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.082303 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.100711 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 19:48:16.094114246 +0000 UTC Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.103127 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:01 crc kubenswrapper[4675]: E0219 08:45:01.103364 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.185526 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.185568 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.185579 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.185598 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.185608 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.289265 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.289325 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.289339 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.289360 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.289371 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.391812 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.391905 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.391915 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.391938 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.391950 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.494883 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.494951 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.494969 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.494993 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.495006 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.598229 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.598281 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.598293 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.598313 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.598327 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.701582 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.701666 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.701677 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.701699 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.701712 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.804951 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.805012 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.805029 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.805054 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.805072 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.907486 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.907532 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.907548 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.907571 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:01 crc kubenswrapper[4675]: I0219 08:45:01.907589 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:01Z","lastTransitionTime":"2026-02-19T08:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.010362 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.010428 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.010441 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.010467 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.010481 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.101300 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 16:23:05.937432351 +0000 UTC Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.102733 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.102749 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.102861 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.103360 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.103703 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.103912 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.113022 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.113087 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.113101 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.113124 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.113142 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.207256 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.207386 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.207424 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207502 4675 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207531 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.207488276 +0000 UTC m=+147.834578544 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207587 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.207575598 +0000 UTC m=+147.834665986 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.207807 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207867 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207935 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207966 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207985 4675 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.207903 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.208054 4675 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.207948 4675 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.208083 4675 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.208069 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.208038602 +0000 UTC m=+147.835128920 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.208109 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.208100854 +0000 UTC m=+147.835191122 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 19 08:45:02 crc kubenswrapper[4675]: E0219 08:45:02.208130 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.208120485 +0000 UTC m=+147.835210743 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.216096 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.216163 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.216181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.216205 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.216217 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.319436 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.319479 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.319493 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.319515 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.319529 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.422640 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.422726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.422739 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.422763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.422776 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.525181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.525249 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.525270 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.525299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.525320 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.628681 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.628763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.628786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.628820 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.628841 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.731865 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.731900 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.731908 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.731922 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.731932 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.834786 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.834865 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.834889 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.834926 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.834949 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.938785 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.938838 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.938848 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.938865 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:02 crc kubenswrapper[4675]: I0219 08:45:02.938875 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:02Z","lastTransitionTime":"2026-02-19T08:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.042462 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.042532 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.042548 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.042572 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.042584 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.102133 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 07:50:47.636095319 +0000 UTC Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.102274 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:03 crc kubenswrapper[4675]: E0219 08:45:03.102459 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.145354 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.145407 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.145418 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.145435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.145445 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.248089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.248147 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.248160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.248211 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.248227 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.350731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.350811 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.350820 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.350840 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.350852 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.453842 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.453901 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.453914 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.453931 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.453942 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.556879 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.556928 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.556939 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.556954 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.556967 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.659778 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.659819 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.659829 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.659848 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.659860 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.762582 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.762619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.762627 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.762642 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.762652 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.864773 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.864809 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.864818 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.864834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.864843 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.967702 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.967766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.967782 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.967809 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:03 crc kubenswrapper[4675]: I0219 08:45:03.967830 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:03Z","lastTransitionTime":"2026-02-19T08:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.070395 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.070443 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.070454 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.070469 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.070481 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.102257 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:04 crc kubenswrapper[4675]: E0219 08:45:04.102406 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.102451 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.102504 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.102442 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 12:09:37.278751649 +0000 UTC Feb 19 08:45:04 crc kubenswrapper[4675]: E0219 08:45:04.102903 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:04 crc kubenswrapper[4675]: E0219 08:45:04.102992 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.173704 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.173782 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.173801 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.173836 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.173854 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.277272 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.277806 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.277818 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.277836 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.277847 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.381310 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.381369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.381382 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.381405 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.381421 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.484577 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.484632 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.484665 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.484686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.484701 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.587910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.587965 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.587980 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.588003 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.588019 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.691029 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.691084 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.691100 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.691121 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.691135 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.794202 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.794246 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.794256 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.794275 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.794287 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.896800 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.896857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.896869 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.896885 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:04 crc kubenswrapper[4675]: I0219 08:45:04.896895 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:04.999907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:04.999976 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:04.999988 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.000010 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.000024 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:04Z","lastTransitionTime":"2026-02-19T08:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.102692 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 02:27:06.108949394 +0000 UTC Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.102718 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.102848 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.102895 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.102908 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.102926 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.102938 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: E0219 08:45:05.102983 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.206300 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.206346 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.206355 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.206372 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.206385 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.310364 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.310436 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.310460 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.310522 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.310538 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.413636 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.413717 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.413729 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.413748 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.413761 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.517134 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.517178 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.517187 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.517205 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.517217 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.619763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.620008 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.620050 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.620083 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.620100 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.722586 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.722626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.722635 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.722671 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.722685 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.825327 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.825380 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.825393 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.825419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.825431 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.928212 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.928255 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.928265 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.928282 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:05 crc kubenswrapper[4675]: I0219 08:45:05.928294 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:05Z","lastTransitionTime":"2026-02-19T08:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.031634 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.031698 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.031708 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.031731 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.031745 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.102558 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.102612 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.102674 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.102924 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.103048 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 19:26:06.356487178 +0000 UTC Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.103119 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.103194 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.114617 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.135461 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.135531 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.135554 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.135576 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.135591 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.238460 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.238511 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.238522 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.238541 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.238553 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.342081 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.342127 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.342138 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.342160 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.342173 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.445102 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.445156 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.445168 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.445190 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.445201 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.547585 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.547641 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.547651 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.547691 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.547704 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.650650 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.650723 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.650732 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.650750 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.650762 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.753541 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.753584 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.753595 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.753675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.753695 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.856958 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.856995 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.857004 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.857019 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.857030 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.904509 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.904558 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.904570 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.904592 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.904605 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.919518 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.924577 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.924630 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.924658 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.924683 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.924699 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.939541 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.944913 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.944959 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.944968 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.944985 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.944997 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.959351 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.965061 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.965106 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.965114 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.965135 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.965147 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.979265 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.983714 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.983745 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.983757 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.983775 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.983788 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.996916 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:06Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:06 crc kubenswrapper[4675]: E0219 08:45:06.997047 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.999236 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.999287 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.999302 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.999325 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:06 crc kubenswrapper[4675]: I0219 08:45:06.999339 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:06Z","lastTransitionTime":"2026-02-19T08:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.102392 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:07 crc kubenswrapper[4675]: E0219 08:45:07.102576 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.102592 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.102627 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.102677 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.102701 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.102717 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.103768 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 05:52:57.520395112 +0000 UTC Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.205406 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.205459 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.205474 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.205495 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.205509 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.308374 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.308509 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.308520 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.308546 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.308557 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.411489 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.411536 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.411550 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.411569 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.411579 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.514433 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.514501 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.514514 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.514539 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.514553 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.618148 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.618218 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.618232 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.618258 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.618275 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.721192 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.721238 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.721247 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.721268 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.721280 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.824170 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.824232 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.824243 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.824265 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.824277 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.927406 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.927481 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.927494 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.927525 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:07 crc kubenswrapper[4675]: I0219 08:45:07.927537 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:07Z","lastTransitionTime":"2026-02-19T08:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.030090 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.030147 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.030159 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.030183 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.030196 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.103073 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.103137 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.103165 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:08 crc kubenswrapper[4675]: E0219 08:45:08.103266 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:08 crc kubenswrapper[4675]: E0219 08:45:08.103682 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:08 crc kubenswrapper[4675]: E0219 08:45:08.103834 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.103915 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 00:35:16.005022633 +0000 UTC Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.104042 4675 scope.go:117] "RemoveContainer" containerID="1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d" Feb 19 08:45:08 crc kubenswrapper[4675]: E0219 08:45:08.104252 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.133484 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.133546 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.133559 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.133580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.133594 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.236630 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.236720 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.236763 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.236794 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.236809 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.339921 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.339978 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.339993 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.340019 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.340039 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.442765 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.442819 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.442830 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.442853 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.442871 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.545855 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.545991 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.546077 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.546107 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.546128 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.649462 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.649524 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.649535 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.649557 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.649570 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.752944 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.753014 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.753023 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.753046 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.753056 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.856245 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.856301 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.856316 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.856341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.856356 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.959902 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.959943 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.959953 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.959969 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:08 crc kubenswrapper[4675]: I0219 08:45:08.959986 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:08Z","lastTransitionTime":"2026-02-19T08:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.062192 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.062264 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.062280 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.062306 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.062322 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.102316 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:09 crc kubenswrapper[4675]: E0219 08:45:09.102514 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.104036 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 05:52:49.999425694 +0000 UTC Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.121208 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.139283 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.156963 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.165589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.165675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.165689 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.165711 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.165755 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.171188 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee1983c5-6d6b-407b-88de-4e3179f6bd33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c65048a1f73a3695505b901ea5ad383c9540482d9c10923472f61fda98f8eae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f143cde834aac64e4b7c9d0a3830884e5d494a735ae31f07c4481845693419b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f143cde834aac64e4b7c9d0a3830884e5d494a735ae31f07c4481845693419b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.191901 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.208910 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.225357 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.242905 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.256668 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.269045 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.269098 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.269112 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.269139 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.269152 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.276761 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:53Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0219 08:44:53.059973 6685 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:53.060252 6685 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0219 08:44:53.060742 6685 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:53.060771 6685 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:53.060815 6685 factory.go:656] Stopping watch factory\\\\nI0219 08:44:53.060815 6685 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:53.060835 6685 handler.go:208] Removed *v1.Node event handler 2\\\\nI0219 08:44:53.062927 6685 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0219 08:44:53.062952 6685 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0219 08:44:53.063021 6685 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:53.063059 6685 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0219 08:44:53.063170 6685 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.288379 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.303429 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.314849 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.328293 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.343114 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.354315 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.365139 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.371407 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.371457 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.371468 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.371486 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.371496 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.375797 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.388506 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:09Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.475076 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.475137 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.475154 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.475176 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.475190 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.577626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.577699 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.577709 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.577744 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.577756 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.680068 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.680126 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.680140 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.680163 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.680177 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.783116 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.783189 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.783203 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.783220 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.783233 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.886012 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.886061 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.886072 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.886089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.886101 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.989171 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.989253 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.989267 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.989283 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:09 crc kubenswrapper[4675]: I0219 08:45:09.989293 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:09Z","lastTransitionTime":"2026-02-19T08:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.092304 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.092369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.092380 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.092401 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.092411 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.103178 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.103221 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.103201 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:10 crc kubenswrapper[4675]: E0219 08:45:10.103380 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:10 crc kubenswrapper[4675]: E0219 08:45:10.103512 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:10 crc kubenswrapper[4675]: E0219 08:45:10.103624 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.105168 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 19:14:07.767537006 +0000 UTC Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.194821 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.194869 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.194879 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.194898 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.194909 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.297366 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.297434 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.297444 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.297461 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.297472 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.400450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.400495 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.400504 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.400524 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.400533 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.503954 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.504004 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.504014 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.504033 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.504044 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.607057 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.607109 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.607119 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.607137 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.607149 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.709659 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.709711 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.709723 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.709744 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.709759 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.812580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.812665 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.812683 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.812708 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.812722 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.916491 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.916544 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.916554 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.916575 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:10 crc kubenswrapper[4675]: I0219 08:45:10.916591 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:10Z","lastTransitionTime":"2026-02-19T08:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.019920 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.019980 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.019992 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.020015 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.020028 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.102461 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:11 crc kubenswrapper[4675]: E0219 08:45:11.102606 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.105379 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 22:33:20.401825596 +0000 UTC Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.123580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.123620 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.123633 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.123670 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.123680 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.226495 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.226557 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.226569 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.226589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.226605 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.329893 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.329951 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.329963 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.329979 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.329988 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.432877 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.432948 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.432958 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.432977 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.432988 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.535734 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.535802 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.535814 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.535828 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.535837 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.638678 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.638753 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.638774 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.638802 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.638832 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.740629 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.740686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.740695 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.740709 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.740720 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.843361 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.843409 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.843418 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.843435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.843445 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.946232 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.946275 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.946286 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.946301 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:11 crc kubenswrapper[4675]: I0219 08:45:11.946310 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:11Z","lastTransitionTime":"2026-02-19T08:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.049550 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.049601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.049612 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.049630 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.049698 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.102950 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.103034 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.103085 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:12 crc kubenswrapper[4675]: E0219 08:45:12.103112 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:12 crc kubenswrapper[4675]: E0219 08:45:12.103316 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:12 crc kubenswrapper[4675]: E0219 08:45:12.103365 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.105886 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 02:55:46.122961217 +0000 UTC Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.152757 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.152809 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.152821 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.152843 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.152855 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.255603 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.255672 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.255687 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.255705 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.255717 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.358435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.358496 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.358508 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.358525 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.358535 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.461972 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.462016 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.462025 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.462041 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.462052 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.565039 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.565114 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.565127 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.565149 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.565164 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.667605 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.667670 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.667680 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.667696 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.667707 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.771217 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.771271 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.771282 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.771303 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.771317 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.873421 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.873505 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.873516 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.873533 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.873544 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.975988 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.976043 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.976060 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.976083 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:12 crc kubenswrapper[4675]: I0219 08:45:12.976097 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:12Z","lastTransitionTime":"2026-02-19T08:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.079899 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.079952 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.079964 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.079983 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.079994 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.102320 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:13 crc kubenswrapper[4675]: E0219 08:45:13.102478 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.106514 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 18:27:06.783377428 +0000 UTC Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.182484 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.182524 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.182534 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.182552 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.182562 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.285485 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.285536 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.285546 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.285563 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.285576 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.388561 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.388669 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.388687 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.388716 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.388735 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.491563 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.491619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.491633 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.491683 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.491697 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.594687 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.594738 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.594748 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.594769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.594782 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.697975 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.698031 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.698042 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.698061 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.698071 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.801793 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.801837 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.801846 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.801867 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.801877 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.905221 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.905283 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.905297 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.905320 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:13 crc kubenswrapper[4675]: I0219 08:45:13.905334 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:13Z","lastTransitionTime":"2026-02-19T08:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.008393 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.008430 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.008443 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.008459 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.008473 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.102521 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.102557 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.102522 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:14 crc kubenswrapper[4675]: E0219 08:45:14.102838 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:14 crc kubenswrapper[4675]: E0219 08:45:14.102995 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:14 crc kubenswrapper[4675]: E0219 08:45:14.103397 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.106725 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 17:52:00.335424519 +0000 UTC Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.111300 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.111374 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.111390 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.111416 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.111432 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.213758 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.213864 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.213884 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.213907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.213927 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.316973 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.317031 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.317043 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.317068 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.317079 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.419687 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.419738 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.419751 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.419766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.419776 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.523768 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.523821 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.523834 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.523853 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.523872 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.626319 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.626381 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.626398 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.626423 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.626439 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.729119 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.729181 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.729196 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.729213 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.729223 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.832531 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.832601 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.832626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.832701 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.832730 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.936412 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.936468 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.936485 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.936511 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:14 crc kubenswrapper[4675]: I0219 08:45:14.936530 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:14Z","lastTransitionTime":"2026-02-19T08:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.038414 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.038457 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.038473 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.038489 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.038499 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.102804 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:15 crc kubenswrapper[4675]: E0219 08:45:15.103050 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.109121 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 23:31:40.238706725 +0000 UTC Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.141421 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.141460 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.141469 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.141484 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.141494 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.244372 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.244424 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.244435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.244456 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.244465 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.347686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.347733 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.347747 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.347766 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.347780 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.450409 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.450469 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.450480 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.450505 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.450519 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.554198 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.554269 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.554286 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.554312 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.554329 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.657125 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.657171 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.657183 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.657203 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.657217 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.760769 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.760830 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.760846 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.760871 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.760883 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.863113 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.863202 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.863218 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.863235 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.863246 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.965413 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.965472 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.965490 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.965514 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:15 crc kubenswrapper[4675]: I0219 08:45:15.965531 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:15Z","lastTransitionTime":"2026-02-19T08:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.068353 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.068394 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.068406 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.068423 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.068434 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.102757 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.102787 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.102757 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:16 crc kubenswrapper[4675]: E0219 08:45:16.102898 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:16 crc kubenswrapper[4675]: E0219 08:45:16.102959 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:16 crc kubenswrapper[4675]: E0219 08:45:16.103042 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.109897 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 13:12:17.247019104 +0000 UTC Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.171507 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.171558 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.171571 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.171589 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.171600 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.275521 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.275570 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.275581 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.275599 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.275612 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.378575 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.378627 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.378650 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.378675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.378687 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.481811 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.481853 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.481864 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.481880 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.481892 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.584219 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.584264 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.584273 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.584291 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.584302 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.687575 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.687686 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.687706 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.687760 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.687780 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.790953 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.791011 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.791022 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.791047 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.791061 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.894219 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.894284 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.894294 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.894309 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.894318 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.996743 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.996788 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.996798 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.996814 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:16 crc kubenswrapper[4675]: I0219 08:45:16.996829 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:16Z","lastTransitionTime":"2026-02-19T08:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.099681 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.099725 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.099735 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.099752 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.099763 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.103230 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:17 crc kubenswrapper[4675]: E0219 08:45:17.103857 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.110161 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 13:56:17.570226298 +0000 UTC Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.202796 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.202863 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.202881 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.202911 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.202931 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.253242 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.253299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.253309 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.253330 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.253342 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: E0219 08:45:17.268300 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:17Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.273852 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.273911 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.273931 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.273961 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.273981 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: E0219 08:45:17.290335 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:17Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.294878 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.294907 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.294916 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.294932 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.294945 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: E0219 08:45:17.310011 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:17Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.315562 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.315654 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.315666 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.315680 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.315690 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: E0219 08:45:17.334481 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:17Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.339282 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.339341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.339351 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.339369 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.339381 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: E0219 08:45:17.353046 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-19T08:45:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"42540fd8-b49d-428b-9ca1-9222fab1c7ed\\\",\\\"systemUUID\\\":\\\"8ed5780f-8c9e-4eed-a721-460b032942bc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:17Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:17 crc kubenswrapper[4675]: E0219 08:45:17.353210 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.355031 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.355059 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.355071 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.355084 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.355095 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.458107 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.458173 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.458190 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.458213 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.458231 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.561099 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.561191 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.561234 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.561295 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.561317 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.665040 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.665095 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.665104 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.665121 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.665130 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.768076 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.768163 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.768177 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.768190 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.768200 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.871796 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.871851 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.871863 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.871882 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.871909 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.975476 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.975525 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.975536 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.975554 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:17 crc kubenswrapper[4675]: I0219 08:45:17.975569 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:17Z","lastTransitionTime":"2026-02-19T08:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.078250 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.078310 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.078321 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.078338 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.078347 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.090912 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:18 crc kubenswrapper[4675]: E0219 08:45:18.091082 4675 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:45:18 crc kubenswrapper[4675]: E0219 08:45:18.091155 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs podName:58bc9d86-8ee5-4635-9de3-23890597f5ca nodeName:}" failed. No retries permitted until 2026-02-19 08:46:22.091132224 +0000 UTC m=+163.718222532 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs") pod "network-metrics-daemon-vcqbz" (UID: "58bc9d86-8ee5-4635-9de3-23890597f5ca") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.102990 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.103077 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.102998 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:18 crc kubenswrapper[4675]: E0219 08:45:18.103181 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:18 crc kubenswrapper[4675]: E0219 08:45:18.103280 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:18 crc kubenswrapper[4675]: E0219 08:45:18.103456 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.111250 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 04:32:45.286283931 +0000 UTC Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.181282 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.181336 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.181349 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.181371 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.181387 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.284975 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.285035 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.285052 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.285083 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.285102 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.388038 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.388114 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.388138 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.388165 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.388185 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.490675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.490728 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.490742 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.490760 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.490769 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.594903 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.594945 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.594956 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.594983 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.594995 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.697707 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.697770 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.697783 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.697805 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.697817 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.801024 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.801074 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.801087 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.801108 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.801120 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.903865 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.903931 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.903942 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.903969 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:18 crc kubenswrapper[4675]: I0219 08:45:18.903985 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:18Z","lastTransitionTime":"2026-02-19T08:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.006453 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.006528 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.006539 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.006567 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.006581 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.102960 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:19 crc kubenswrapper[4675]: E0219 08:45:19.103141 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.110548 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.110623 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.110671 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.110708 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.110736 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.111376 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 00:15:27.694590324 +0000 UTC Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.127754 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7912463c-4c5e-4269-a911-814275e128b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64e4a0486ab1c11eb85b63b73f795c217d959fcb96a40e635004d70b427a6b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05cbbd302c9f69a0cea8c998d851c452044301abee085220ecd6deadeacb9308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d573511109eff9e12e144eb53da0ab366ceb1afb1420686f014ef2e12001225\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bc713ea42f062dc194fbe3782a8fcf00effe61715b57b04b8543300b3a89ff3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c44d0f3e1c4b4ca96296218938dfc318f712f66c5a7f49532c581b8f42909ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b6dfdfb5b73a49d99c980f362525a4c3f8534285ac2f27e6b763e106f6b9e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452656f378fe88885268d96cb4ca97ce9e2597c6458e6b79dc1868414584ffbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrmjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h6jd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.148733 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8a735fe-d7b1-461c-ab2b-81cbc9f687fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30c3f1cc60022be00fd16fb74021dd522631dae842906822607e9fe4e05c6003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a42f1f303b5fea5d945c37a515ce2f442f4bb35ec9587c83fcd30440b282aa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed951a7ee54bc3e8ec00584f8c123105deb95ea082f0016d816e47f93b13f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://409355a511e54ae4d95d705ad79527c08f0b756bb457b5c35397b1f07c857be6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.170122 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.192180 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hj427" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92c6f02-3c18-4d7a-b559-840c8415db58\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44cb1c9be669550fb7cd560b3b6fd56afaaade33e2b33611c73f20c208531061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjb8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hj427\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.211150 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e91469d-12b7-434c-991d-633287712a69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a2d43680cce26fe8b57fc01afe07ade2e077adac87e2420d1c2453f4d5bb2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4j22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wqmkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.212425 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.212448 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.212455 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.212470 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.212481 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.230876 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qmz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cc07c5a-23e4-478f-88a7-73d41e17a320\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:48Z\\\",\\\"message\\\":\\\"2026-02-19T08:44:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9\\\\n2026-02-19T08:44:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_71a2360d-959a-402c-8399-74cf6c04f0f9 to /host/opt/cni/bin/\\\\n2026-02-19T08:44:03Z [verbose] multus-daemon started\\\\n2026-02-19T08:44:03Z [verbose] Readiness Indicator file check\\\\n2026-02-19T08:44:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzdkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qmz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.248060 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.266320 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.281533 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e091925247b568018c9a6fee392ee3745866ebdebdf43a4ec86e6d277e602e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8643c32ece9120813887058242f51524a197f43d34f860cad9c08606c73693db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.299235 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee1983c5-6d6b-407b-88de-4e3179f6bd33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c65048a1f73a3695505b901ea5ad383c9540482d9c10923472f61fda98f8eae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f143cde834aac64e4b7c9d0a3830884e5d494a735ae31f07c4481845693419b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f143cde834aac64e4b7c9d0a3830884e5d494a735ae31f07c4481845693419b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.315224 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.315284 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.315299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.315318 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.315331 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.327691 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0143cf58-7024-4233-b0ae-54828d7618d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e47a6c2a5c3cbc2d0b6bdaa987f54f6a21efbd8ff04da2839d0726bc1f312a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6afb24427976786950904f7603bf85ba2199239cdd986edbdacb95593265fa36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ebcf5eafea5eadcc89fbec6c5845831d6835cb209d4a47c61cf635bdd7c501e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://295334dfe6282c609964bc4ffd47711dad7a27ab1b0439d081cc1f724203a5d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a712cad4eae971e7026657762cee357f79efd6df9515cbd33aa96584b3cb05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://817cecea886fa62cf3877a2eeab80fe9e2177d178941e0143267471c84f29cdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9599922c7c47615c61c3843443db035751368a7e8b3579fc487f22d16ebd007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c86c0383d1eafd1b1b0a6539ecac55a0034431e53771652535a7ba5f082172f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.343812 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffd5e4b9-c1df-47cf-bf03-8596cdbd375e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-19T08:43:58Z\\\",\\\"message\\\":\\\" 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771490622 cert, and key in /tmp/serving-cert-4138014238/serving-signer.crt, /tmp/serving-cert-4138014238/serving-signer.key\\\\nI0219 08:43:42.800287 1 observer_polling.go:159] Starting file observer\\\\nW0219 08:43:52.804231 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0219 08:43:52.804425 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0219 08:43:52.807545 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4138014238/tls.crt::/tmp/serving-cert-4138014238/tls.key\\\\\\\"\\\\nI0219 08:43:58.221302 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0219 08:43:58.224340 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0219 08:43:58.224363 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0219 08:43:58.224389 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0219 08:43:58.224396 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nF0219 08:43:58.243360 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0219 08:43:58.245800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0219 08:43:58.245846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0219 08:43:58.245875 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245947 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0219 08:43:58.245981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:43:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.357027 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bad83622-dfaf-441a-97eb-91b249a4e4db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d2c6ca84e13630706e62ff3096492a110d2abd94fccbead046307978769fcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee8a10180026cb31ac6d7c9df01d599a469c953d768e02bb3d3414321c84e95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68cbd1013902881da2ef7b0992daf3eb207757c3da5754b4a72b33e22c6ef7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.369967 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d26f9d7485c470d5bbb7cde9e497ae9853556ade9dc7fc7f937c53138200f7df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.385717 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d52834d2a6b4f332dffe5d428c35a5e53891ed8ed3929dbadba2c6a42fbb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.405681 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d95acb1e-cd60-48dd-af45-e190cd07723a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-19T08:44:53Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0219 08:44:53.059973 6685 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0219 08:44:53.060252 6685 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0219 08:44:53.060742 6685 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0219 08:44:53.060771 6685 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0219 08:44:53.060815 6685 factory.go:656] Stopping watch factory\\\\nI0219 08:44:53.060815 6685 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0219 08:44:53.060835 6685 handler.go:208] Removed *v1.Node event handler 2\\\\nI0219 08:44:53.062927 6685 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0219 08:44:53.062952 6685 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0219 08:44:53.063021 6685 ovnkube.go:599] Stopped ovnkube\\\\nI0219 08:44:53.063059 6685 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0219 08:44:53.063170 6685 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-19T08:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5ndr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lkvd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.420041 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.420174 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.420192 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.420219 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.420324 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6kmw9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c50286d-89aa-4d1c-a424-5fd45d0355fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c92cb21a7e88777f5e27f5498bf7ace2ee272393a323ecc8b6d5549f14c52a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k2frs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:43:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6kmw9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.420325 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.434332 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18ae2562-beb3-4423-9045-f8559d492d8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22f87a6bc5b751d37d06d01106c0eefee33b590e25c9f50dc3c66e0747c514db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26562962de1077a2ed3dab967d7110127d1ebad0109ac26dc219945d40b3ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-19T08:44:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tj8nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x2w85\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.447781 4675 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bc9d86-8ee5-4635-9de3-23890597f5ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-19T08:44:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xr6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-19T08:44:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vcqbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-19T08:45:19Z is after 2025-08-24T17:21:41Z" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.524597 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.524694 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.524713 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.524740 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.524755 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.626837 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.626872 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.626881 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.626899 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.626910 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.729882 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.729937 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.729954 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.729982 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.729999 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.833026 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.833069 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.833080 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.833098 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.833112 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.936484 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.936534 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.936543 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.936562 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:19 crc kubenswrapper[4675]: I0219 08:45:19.936574 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:19Z","lastTransitionTime":"2026-02-19T08:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.039311 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.039388 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.039412 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.039441 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.039458 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.103183 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.103220 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.103375 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:20 crc kubenswrapper[4675]: E0219 08:45:20.103494 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:20 crc kubenswrapper[4675]: E0219 08:45:20.103697 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:20 crc kubenswrapper[4675]: E0219 08:45:20.103785 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.111670 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 23:43:33.849464774 +0000 UTC Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.142280 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.142330 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.142341 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.142362 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.142375 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.245566 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.245611 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.245619 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.245650 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.245660 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.348359 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.348399 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.348406 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.348421 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.348432 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.451092 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.451142 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.451150 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.451166 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.451177 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.553653 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.553711 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.553723 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.553740 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.553752 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.656618 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.656690 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.656701 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.656725 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.656736 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.759626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.759700 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.759711 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.759728 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.759738 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.862596 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.862674 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.862685 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.862705 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.862717 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.965922 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.965985 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.965999 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.966026 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:20 crc kubenswrapper[4675]: I0219 08:45:20.966040 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:20Z","lastTransitionTime":"2026-02-19T08:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.069499 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.069557 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.069574 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.069598 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.069615 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.102360 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:21 crc kubenswrapper[4675]: E0219 08:45:21.102853 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.112320 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 07:04:16.279421518 +0000 UTC Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.171909 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.171950 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.171961 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.171980 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.171991 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.275057 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.275116 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.275129 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.275155 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.275170 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.379220 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.379276 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.379286 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.379308 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.379319 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.482877 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.482966 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.482987 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.483013 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.483035 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.585877 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.585926 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.585938 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.585963 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.585978 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.688562 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.688618 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.688631 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.688681 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.688698 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.791348 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.791408 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.791426 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.791450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.791467 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.894733 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.894807 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.894830 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.894857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.894878 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.997918 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.997949 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.997957 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.997970 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:21 crc kubenswrapper[4675]: I0219 08:45:21.997979 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:21Z","lastTransitionTime":"2026-02-19T08:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.100770 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.100833 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.100849 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.100877 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.100898 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.102950 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.102976 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:22 crc kubenswrapper[4675]: E0219 08:45:22.103063 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.103070 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:22 crc kubenswrapper[4675]: E0219 08:45:22.103183 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:22 crc kubenswrapper[4675]: E0219 08:45:22.103221 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.112588 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 08:08:14.430922394 +0000 UTC Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.203675 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.203724 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.203734 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.203749 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.203759 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.307011 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.307057 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.307069 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.307089 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.307103 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.410513 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.410555 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.410571 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.410593 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.410610 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.512937 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.513015 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.513110 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.513142 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.513162 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.626050 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.626122 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.626190 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.626217 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.626232 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.728854 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.729016 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.729042 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.729135 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.729156 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.831998 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.832131 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.832143 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.832157 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.832168 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.935419 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.935502 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.935513 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.935533 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:22 crc kubenswrapper[4675]: I0219 08:45:22.935543 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:22Z","lastTransitionTime":"2026-02-19T08:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.038819 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.038878 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.038889 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.038910 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.038922 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.103197 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:23 crc kubenswrapper[4675]: E0219 08:45:23.103442 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.104953 4675 scope.go:117] "RemoveContainer" containerID="1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d" Feb 19 08:45:23 crc kubenswrapper[4675]: E0219 08:45:23.105310 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lkvd2_openshift-ovn-kubernetes(d95acb1e-cd60-48dd-af45-e190cd07723a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.113026 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 21:55:34.465526328 +0000 UTC Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.140859 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.140916 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.140927 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.140944 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.140957 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.243260 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.243299 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.243310 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.243324 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.243335 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.346946 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.347004 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.347021 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.347059 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.347075 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.450497 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.450563 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.450580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.450604 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.450622 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.553411 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.553466 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.553484 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.553508 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.553525 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.656252 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.656304 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.656313 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.656327 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.656357 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.759005 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.759044 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.759055 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.759071 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.759082 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.861796 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.861861 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.861879 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.861904 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.861923 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.964959 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.965022 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.965043 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.965069 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:23 crc kubenswrapper[4675]: I0219 08:45:23.965089 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:23Z","lastTransitionTime":"2026-02-19T08:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.068708 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.068768 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.068787 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.068814 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.068831 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.102565 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.102610 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:24 crc kubenswrapper[4675]: E0219 08:45:24.102725 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.102748 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:24 crc kubenswrapper[4675]: E0219 08:45:24.103041 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:24 crc kubenswrapper[4675]: E0219 08:45:24.103255 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.114085 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 21:48:07.742834878 +0000 UTC Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.172112 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.172170 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.172186 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.172214 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.172233 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.275413 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.275451 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.275461 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.275475 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.275484 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.378523 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.378580 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.378591 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.378608 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.378622 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.481653 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.481694 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.481703 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.481718 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.481729 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.584033 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.584092 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.584108 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.584132 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.584148 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.686661 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.686696 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.686706 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.686724 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.686734 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.790558 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.790612 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.790622 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.790660 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.790676 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.892726 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.892803 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.892826 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.892857 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.892880 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.995487 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.995564 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.995577 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.995594 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:24 crc kubenswrapper[4675]: I0219 08:45:24.995604 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:24Z","lastTransitionTime":"2026-02-19T08:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.098333 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.098450 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.098467 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.098495 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.098509 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.102739 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:25 crc kubenswrapper[4675]: E0219 08:45:25.102942 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.114493 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 21:40:06.275508784 +0000 UTC Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.201248 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.201302 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.201314 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.201335 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.201350 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.309584 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.309657 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.309672 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.309698 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.309711 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.412903 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.412957 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.412969 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.412991 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.413007 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.515267 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.515315 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.515325 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.515345 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.515356 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.617547 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.617599 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.617608 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.617626 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.617678 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.720629 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.720698 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.720709 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.720725 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.720736 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.823543 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.823736 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.823816 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.823843 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.823897 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.925811 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.925865 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.925876 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.925900 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:25 crc kubenswrapper[4675]: I0219 08:45:25.925912 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:25Z","lastTransitionTime":"2026-02-19T08:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.029118 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.029166 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.029175 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.029193 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.029203 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.102395 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.102495 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:26 crc kubenswrapper[4675]: E0219 08:45:26.102572 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.102599 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:26 crc kubenswrapper[4675]: E0219 08:45:26.102768 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:26 crc kubenswrapper[4675]: E0219 08:45:26.102870 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.115672 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:48:19.652126273 +0000 UTC Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.131957 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.132014 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.132025 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.132046 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.132060 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.234362 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.234420 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.234435 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.234454 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.234469 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.345216 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.345266 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.345283 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.345306 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.345326 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.448387 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.448444 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.448464 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.448489 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.448508 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.558593 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.558632 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.558666 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.558685 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.558697 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.661280 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.661350 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.661371 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.661403 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.661426 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.763582 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.763690 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.763700 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.763723 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.763739 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.866406 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.866465 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.866477 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.866496 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.866513 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.969068 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.969120 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.969128 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.969143 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:26 crc kubenswrapper[4675]: I0219 08:45:26.969169 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:26Z","lastTransitionTime":"2026-02-19T08:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.070992 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.071034 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.071042 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.071058 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.071068 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.102820 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:27 crc kubenswrapper[4675]: E0219 08:45:27.103028 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.116042 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 17:17:32.81908281 +0000 UTC Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.175892 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.176003 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.176025 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.176057 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.176071 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.279448 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.279522 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.279559 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.279590 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.279612 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.382777 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.382868 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.382894 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.382989 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.383017 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.486156 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.486259 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.486279 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.486307 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.486329 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.589538 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.589616 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.589627 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.589661 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.589671 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.692782 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.692846 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.692864 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.692890 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.692907 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.708047 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.708083 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.708094 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.708109 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.708120 4675 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-19T08:45:27Z","lastTransitionTime":"2026-02-19T08:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.780571 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr"] Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.781467 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.784027 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.784377 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.784614 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.784913 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.850281 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-6kmw9" podStartSLOduration=88.850252244 podStartE2EDuration="1m28.850252244s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.850081729 +0000 UTC m=+109.477172007" watchObservedRunningTime="2026-02-19 08:45:27.850252244 +0000 UTC m=+109.477342512" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.868453 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x2w85" podStartSLOduration=88.868423263 podStartE2EDuration="1m28.868423263s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.865749693 +0000 UTC m=+109.492839991" watchObservedRunningTime="2026-02-19 08:45:27.868423263 +0000 UTC m=+109.495513571" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.892145 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=58.892117936 podStartE2EDuration="58.892117936s" podCreationTimestamp="2026-02-19 08:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.891490087 +0000 UTC m=+109.518580355" watchObservedRunningTime="2026-02-19 08:45:27.892117936 +0000 UTC m=+109.519208214" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.900320 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7613ecc8-daa6-4145-9589-fd8e83e40444-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.900370 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7613ecc8-daa6-4145-9589-fd8e83e40444-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.900398 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7613ecc8-daa6-4145-9589-fd8e83e40444-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.900427 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7613ecc8-daa6-4145-9589-fd8e83e40444-service-ca\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.900506 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7613ecc8-daa6-4145-9589-fd8e83e40444-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.946043 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podStartSLOduration=88.946024635 podStartE2EDuration="1m28.946024635s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.945701315 +0000 UTC m=+109.572791583" watchObservedRunningTime="2026-02-19 08:45:27.946024635 +0000 UTC m=+109.573114903" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.946689 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hj427" podStartSLOduration=88.946681564 podStartE2EDuration="1m28.946681564s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.922353443 +0000 UTC m=+109.549443711" watchObservedRunningTime="2026-02-19 08:45:27.946681564 +0000 UTC m=+109.573771822" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.967004 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-qmz2s" podStartSLOduration=88.966991267 podStartE2EDuration="1m28.966991267s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.966849773 +0000 UTC m=+109.593940041" watchObservedRunningTime="2026-02-19 08:45:27.966991267 +0000 UTC m=+109.594081535" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.986353 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-h6jd6" podStartSLOduration=88.986332431 podStartE2EDuration="1m28.986332431s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.985571008 +0000 UTC m=+109.612661276" watchObservedRunningTime="2026-02-19 08:45:27.986332431 +0000 UTC m=+109.613422699" Feb 19 08:45:27 crc kubenswrapper[4675]: I0219 08:45:27.998153 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.998129671 podStartE2EDuration="21.998129671s" podCreationTimestamp="2026-02-19 08:45:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:27.997285546 +0000 UTC m=+109.624375814" watchObservedRunningTime="2026-02-19 08:45:27.998129671 +0000 UTC m=+109.625219939" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.001151 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7613ecc8-daa6-4145-9589-fd8e83e40444-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.001522 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7613ecc8-daa6-4145-9589-fd8e83e40444-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.001683 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7613ecc8-daa6-4145-9589-fd8e83e40444-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.001888 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7613ecc8-daa6-4145-9589-fd8e83e40444-service-ca\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.002867 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7613ecc8-daa6-4145-9589-fd8e83e40444-service-ca\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.002922 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7613ecc8-daa6-4145-9589-fd8e83e40444-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.003015 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7613ecc8-daa6-4145-9589-fd8e83e40444-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.001845 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7613ecc8-daa6-4145-9589-fd8e83e40444-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.010530 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7613ecc8-daa6-4145-9589-fd8e83e40444-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.019524 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7613ecc8-daa6-4145-9589-fd8e83e40444-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-85hvr\" (UID: \"7613ecc8-daa6-4145-9589-fd8e83e40444\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.026951 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=86.026930475 podStartE2EDuration="1m26.026930475s" podCreationTimestamp="2026-02-19 08:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:28.026923905 +0000 UTC m=+109.654014173" watchObservedRunningTime="2026-02-19 08:45:28.026930475 +0000 UTC m=+109.654020743" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.062626 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.062600813 podStartE2EDuration="1m28.062600813s" podCreationTimestamp="2026-02-19 08:44:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:28.062423598 +0000 UTC m=+109.689513866" watchObservedRunningTime="2026-02-19 08:45:28.062600813 +0000 UTC m=+109.689691081" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.063383 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.063376206 podStartE2EDuration="1m29.063376206s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:28.04598019 +0000 UTC m=+109.673070458" watchObservedRunningTime="2026-02-19 08:45:28.063376206 +0000 UTC m=+109.690466474" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.103041 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:28 crc kubenswrapper[4675]: E0219 08:45:28.103219 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.103574 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:28 crc kubenswrapper[4675]: E0219 08:45:28.103692 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.103882 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.104746 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:28 crc kubenswrapper[4675]: E0219 08:45:28.105058 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.116570 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 04:52:44.188560727 +0000 UTC Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.117182 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.144057 4675 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.930373 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" event={"ID":"7613ecc8-daa6-4145-9589-fd8e83e40444","Type":"ContainerStarted","Data":"a57ccc8d7da4c756296a4dd7efd9d7541effaa5d518723638e06f8f643e272b7"} Feb 19 08:45:28 crc kubenswrapper[4675]: I0219 08:45:28.930884 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" event={"ID":"7613ecc8-daa6-4145-9589-fd8e83e40444","Type":"ContainerStarted","Data":"de328a770354e26d7f3f9f85b56e5ff2738497eabd2636aef516f609e6c16d35"} Feb 19 08:45:29 crc kubenswrapper[4675]: I0219 08:45:29.102593 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:29 crc kubenswrapper[4675]: E0219 08:45:29.103763 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:30 crc kubenswrapper[4675]: I0219 08:45:30.102704 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:30 crc kubenswrapper[4675]: I0219 08:45:30.102819 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:30 crc kubenswrapper[4675]: I0219 08:45:30.102865 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:30 crc kubenswrapper[4675]: E0219 08:45:30.102954 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:30 crc kubenswrapper[4675]: E0219 08:45:30.103197 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:30 crc kubenswrapper[4675]: E0219 08:45:30.103250 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:31 crc kubenswrapper[4675]: I0219 08:45:31.102451 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:31 crc kubenswrapper[4675]: E0219 08:45:31.103256 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:32 crc kubenswrapper[4675]: I0219 08:45:32.103131 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:32 crc kubenswrapper[4675]: I0219 08:45:32.103294 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:32 crc kubenswrapper[4675]: E0219 08:45:32.103491 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:32 crc kubenswrapper[4675]: E0219 08:45:32.103305 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:32 crc kubenswrapper[4675]: I0219 08:45:32.104355 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:32 crc kubenswrapper[4675]: E0219 08:45:32.104705 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:33 crc kubenswrapper[4675]: I0219 08:45:33.102765 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:33 crc kubenswrapper[4675]: E0219 08:45:33.103008 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:34 crc kubenswrapper[4675]: I0219 08:45:34.102289 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:34 crc kubenswrapper[4675]: I0219 08:45:34.102361 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:34 crc kubenswrapper[4675]: I0219 08:45:34.102415 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:34 crc kubenswrapper[4675]: E0219 08:45:34.102498 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:34 crc kubenswrapper[4675]: E0219 08:45:34.102733 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:34 crc kubenswrapper[4675]: E0219 08:45:34.102872 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:35 crc kubenswrapper[4675]: I0219 08:45:35.103227 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:35 crc kubenswrapper[4675]: E0219 08:45:35.103470 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:35 crc kubenswrapper[4675]: I0219 08:45:35.955605 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/1.log" Feb 19 08:45:35 crc kubenswrapper[4675]: I0219 08:45:35.956084 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/0.log" Feb 19 08:45:35 crc kubenswrapper[4675]: I0219 08:45:35.956142 4675 generic.go:334] "Generic (PLEG): container finished" podID="4cc07c5a-23e4-478f-88a7-73d41e17a320" containerID="413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064" exitCode=1 Feb 19 08:45:35 crc kubenswrapper[4675]: I0219 08:45:35.956185 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerDied","Data":"413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064"} Feb 19 08:45:35 crc kubenswrapper[4675]: I0219 08:45:35.956232 4675 scope.go:117] "RemoveContainer" containerID="b63a45429abf6e7b08ff6eebdc402007189ae65155195a8989d75c8dd790261b" Feb 19 08:45:35 crc kubenswrapper[4675]: I0219 08:45:35.957070 4675 scope.go:117] "RemoveContainer" containerID="413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064" Feb 19 08:45:35 crc kubenswrapper[4675]: E0219 08:45:35.957367 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-qmz2s_openshift-multus(4cc07c5a-23e4-478f-88a7-73d41e17a320)\"" pod="openshift-multus/multus-qmz2s" podUID="4cc07c5a-23e4-478f-88a7-73d41e17a320" Feb 19 08:45:36 crc kubenswrapper[4675]: I0219 08:45:36.007591 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-85hvr" podStartSLOduration=97.007553018 podStartE2EDuration="1m37.007553018s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:28.947601917 +0000 UTC m=+110.574692185" watchObservedRunningTime="2026-02-19 08:45:36.007553018 +0000 UTC m=+117.634643306" Feb 19 08:45:36 crc kubenswrapper[4675]: I0219 08:45:36.102618 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:36 crc kubenswrapper[4675]: I0219 08:45:36.102674 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:36 crc kubenswrapper[4675]: I0219 08:45:36.102677 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:36 crc kubenswrapper[4675]: E0219 08:45:36.102813 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:36 crc kubenswrapper[4675]: E0219 08:45:36.102906 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:36 crc kubenswrapper[4675]: E0219 08:45:36.102990 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:36 crc kubenswrapper[4675]: I0219 08:45:36.967189 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/1.log" Feb 19 08:45:37 crc kubenswrapper[4675]: I0219 08:45:37.103349 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:37 crc kubenswrapper[4675]: E0219 08:45:37.103534 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.102496 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.102574 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.102628 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:38 crc kubenswrapper[4675]: E0219 08:45:38.102671 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:38 crc kubenswrapper[4675]: E0219 08:45:38.102807 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:38 crc kubenswrapper[4675]: E0219 08:45:38.102977 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.103781 4675 scope.go:117] "RemoveContainer" containerID="1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d" Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.966409 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vcqbz"] Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.982571 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/3.log" Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.984976 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerStarted","Data":"0a39b552dc1baa97f2a5ddf350fead6cbe122845f858e49017a916ed0ee17b36"} Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.985031 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:38 crc kubenswrapper[4675]: E0219 08:45:38.985160 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:38 crc kubenswrapper[4675]: I0219 08:45:38.985355 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:45:39 crc kubenswrapper[4675]: I0219 08:45:39.010085 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podStartSLOduration=100.010063226 podStartE2EDuration="1m40.010063226s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:45:39.009334774 +0000 UTC m=+120.636425042" watchObservedRunningTime="2026-02-19 08:45:39.010063226 +0000 UTC m=+120.637153494" Feb 19 08:45:39 crc kubenswrapper[4675]: E0219 08:45:39.035336 4675 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 19 08:45:39 crc kubenswrapper[4675]: I0219 08:45:39.103995 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:39 crc kubenswrapper[4675]: E0219 08:45:39.104506 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:39 crc kubenswrapper[4675]: E0219 08:45:39.207277 4675 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 19 08:45:40 crc kubenswrapper[4675]: I0219 08:45:40.102529 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:40 crc kubenswrapper[4675]: I0219 08:45:40.102615 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:40 crc kubenswrapper[4675]: E0219 08:45:40.102705 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:40 crc kubenswrapper[4675]: E0219 08:45:40.102850 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:41 crc kubenswrapper[4675]: I0219 08:45:41.103000 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:41 crc kubenswrapper[4675]: I0219 08:45:41.103000 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:41 crc kubenswrapper[4675]: E0219 08:45:41.103593 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:41 crc kubenswrapper[4675]: E0219 08:45:41.103664 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:42 crc kubenswrapper[4675]: I0219 08:45:42.103212 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:42 crc kubenswrapper[4675]: I0219 08:45:42.103262 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:42 crc kubenswrapper[4675]: E0219 08:45:42.103385 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:42 crc kubenswrapper[4675]: E0219 08:45:42.103571 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:43 crc kubenswrapper[4675]: I0219 08:45:43.102602 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:43 crc kubenswrapper[4675]: I0219 08:45:43.102602 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:43 crc kubenswrapper[4675]: E0219 08:45:43.102788 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:43 crc kubenswrapper[4675]: E0219 08:45:43.102844 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:44 crc kubenswrapper[4675]: I0219 08:45:44.102407 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:44 crc kubenswrapper[4675]: I0219 08:45:44.102492 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:44 crc kubenswrapper[4675]: E0219 08:45:44.102561 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:44 crc kubenswrapper[4675]: E0219 08:45:44.102706 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:44 crc kubenswrapper[4675]: E0219 08:45:44.209091 4675 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 19 08:45:45 crc kubenswrapper[4675]: I0219 08:45:45.103232 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:45 crc kubenswrapper[4675]: I0219 08:45:45.103308 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:45 crc kubenswrapper[4675]: E0219 08:45:45.103426 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:45 crc kubenswrapper[4675]: E0219 08:45:45.103577 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:46 crc kubenswrapper[4675]: I0219 08:45:46.102689 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:46 crc kubenswrapper[4675]: I0219 08:45:46.102791 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:46 crc kubenswrapper[4675]: E0219 08:45:46.103172 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:46 crc kubenswrapper[4675]: E0219 08:45:46.103349 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:47 crc kubenswrapper[4675]: I0219 08:45:47.103363 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:47 crc kubenswrapper[4675]: I0219 08:45:47.103377 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:47 crc kubenswrapper[4675]: E0219 08:45:47.103578 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:47 crc kubenswrapper[4675]: E0219 08:45:47.103732 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:48 crc kubenswrapper[4675]: I0219 08:45:48.102612 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:48 crc kubenswrapper[4675]: I0219 08:45:48.102677 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:48 crc kubenswrapper[4675]: E0219 08:45:48.102798 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:48 crc kubenswrapper[4675]: E0219 08:45:48.102953 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:49 crc kubenswrapper[4675]: I0219 08:45:49.103001 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:49 crc kubenswrapper[4675]: E0219 08:45:49.104868 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:49 crc kubenswrapper[4675]: I0219 08:45:49.104970 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:49 crc kubenswrapper[4675]: E0219 08:45:49.105158 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:49 crc kubenswrapper[4675]: E0219 08:45:49.209707 4675 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 19 08:45:50 crc kubenswrapper[4675]: I0219 08:45:50.102446 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:50 crc kubenswrapper[4675]: I0219 08:45:50.102465 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:50 crc kubenswrapper[4675]: E0219 08:45:50.102748 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:50 crc kubenswrapper[4675]: E0219 08:45:50.102855 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:51 crc kubenswrapper[4675]: I0219 08:45:51.102783 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:51 crc kubenswrapper[4675]: E0219 08:45:51.103001 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:51 crc kubenswrapper[4675]: I0219 08:45:51.103396 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:51 crc kubenswrapper[4675]: I0219 08:45:51.103393 4675 scope.go:117] "RemoveContainer" containerID="413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064" Feb 19 08:45:51 crc kubenswrapper[4675]: E0219 08:45:51.103582 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:52 crc kubenswrapper[4675]: I0219 08:45:52.044280 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/1.log" Feb 19 08:45:52 crc kubenswrapper[4675]: I0219 08:45:52.044724 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerStarted","Data":"664fc053ea88093ccbbf46c3c42c961a4cfe2c8ce85fca2abe20390388336078"} Feb 19 08:45:52 crc kubenswrapper[4675]: I0219 08:45:52.103231 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:52 crc kubenswrapper[4675]: E0219 08:45:52.103394 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:52 crc kubenswrapper[4675]: I0219 08:45:52.106896 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:52 crc kubenswrapper[4675]: E0219 08:45:52.107256 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:53 crc kubenswrapper[4675]: I0219 08:45:53.102705 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:53 crc kubenswrapper[4675]: I0219 08:45:53.102721 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:53 crc kubenswrapper[4675]: E0219 08:45:53.103581 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 19 08:45:53 crc kubenswrapper[4675]: E0219 08:45:53.103962 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vcqbz" podUID="58bc9d86-8ee5-4635-9de3-23890597f5ca" Feb 19 08:45:54 crc kubenswrapper[4675]: I0219 08:45:54.102456 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:54 crc kubenswrapper[4675]: I0219 08:45:54.102490 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:54 crc kubenswrapper[4675]: E0219 08:45:54.102675 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 19 08:45:54 crc kubenswrapper[4675]: E0219 08:45:54.102851 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 19 08:45:55 crc kubenswrapper[4675]: I0219 08:45:55.102316 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:45:55 crc kubenswrapper[4675]: I0219 08:45:55.102443 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:45:55 crc kubenswrapper[4675]: I0219 08:45:55.106686 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 19 08:45:55 crc kubenswrapper[4675]: I0219 08:45:55.106812 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 19 08:45:55 crc kubenswrapper[4675]: I0219 08:45:55.107014 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 19 08:45:55 crc kubenswrapper[4675]: I0219 08:45:55.107081 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 19 08:45:56 crc kubenswrapper[4675]: I0219 08:45:56.102916 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:45:56 crc kubenswrapper[4675]: I0219 08:45:56.102910 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:45:56 crc kubenswrapper[4675]: I0219 08:45:56.105562 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 19 08:45:56 crc kubenswrapper[4675]: I0219 08:45:56.106744 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.219076 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.532789 4675 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.567383 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6nnjw"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.567849 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.569298 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.571097 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.572101 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hhkkj"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.577720 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-pr5cf"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.578081 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-pr5cf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.578698 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.578902 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.579365 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.579450 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.579686 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.579681 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.579860 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rt65v"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.580410 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.580959 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.581286 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.581841 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.581946 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.582065 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-plmm7"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.582583 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.583256 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.583326 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.583370 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.583455 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.583692 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.583779 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.584224 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.584239 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.584735 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.585108 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.585686 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q6tbj"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.586440 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.592707 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.592996 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.593141 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.593581 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vg9l6"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.594354 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.594490 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.594604 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.594748 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.594758 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sdrxk"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.594868 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.595013 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.595128 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.595263 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.595285 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.595398 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.595521 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.595604 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.596263 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.596486 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.596507 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.596830 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.597589 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.597606 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.597830 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.597874 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.597929 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.599384 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-srrlc"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.599808 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.599970 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.600612 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.601306 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.601505 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rc9fc"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.602004 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.602394 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.611947 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.612783 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.613075 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.613295 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.613576 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.614403 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.614876 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.615117 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.615452 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.616745 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617082 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617167 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617260 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617430 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617453 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617717 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617885 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.617086 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.618215 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.618562 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.619226 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.619380 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.619418 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.619606 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.619890 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.620050 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jfck7"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.620065 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.620233 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.620721 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6nnjw"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.620832 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.623534 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-pr5cf"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.624986 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hhkkj"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.626990 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.627230 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.632311 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.632583 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q6tbj"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.635746 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.643951 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xrzvw"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.644123 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.645386 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.651264 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.652039 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.668526 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.669228 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.669633 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.678314 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679313 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4612eb92-6c6b-457b-bab4-da111cac6428-serving-cert\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679393 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-dir\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679442 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4wtk\" (UniqueName: \"kubernetes.io/projected/cc0c9cff-8b28-4d9c-b50d-0ab81e16c841-kube-api-access-d4wtk\") pod \"downloads-7954f5f757-pr5cf\" (UID: \"cc0c9cff-8b28-4d9c-b50d-0ab81e16c841\") " pod="openshift-console/downloads-7954f5f757-pr5cf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679473 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-config\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679501 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cacb4d41-218a-441d-a62a-77338425aa3e-node-pullsecrets\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679526 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679573 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-serving-cert\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679597 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679622 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-service-ca\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679663 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679685 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679705 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679728 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9feab1ba-d666-44c5-b929-83a1732c8014-auth-proxy-config\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679750 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-policies\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679777 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbr8s\" (UniqueName: \"kubernetes.io/projected/4d1c19bc-9031-49eb-88a1-3cb263134b3d-kube-api-access-mbr8s\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679798 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-oauth-serving-cert\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679824 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twb6k\" (UniqueName: \"kubernetes.io/projected/9feab1ba-d666-44c5-b929-83a1732c8014-kube-api-access-twb6k\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679846 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679871 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2198f599-a3a9-472a-b63e-40e302516c64-serving-cert\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679895 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcnx9\" (UniqueName: \"kubernetes.io/projected/2198f599-a3a9-472a-b63e-40e302516c64-kube-api-access-rcnx9\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679921 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679942 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-serving-cert\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679963 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-audit\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.679989 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680020 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4796\" (UniqueName: \"kubernetes.io/projected/366baa67-abf1-485c-9430-1a891c9e58c7-kube-api-access-s4796\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680044 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3c6ce409-7798-42c3-8f33-f7d2272bdc86-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680068 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrzqz\" (UniqueName: \"kubernetes.io/projected/0b433b98-d626-4286-a89b-3618ba63b58a-kube-api-access-nrzqz\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680092 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5pqc\" (UniqueName: \"kubernetes.io/projected/4612eb92-6c6b-457b-bab4-da111cac6428-kube-api-access-b5pqc\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680116 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-trusted-ca-bundle\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680141 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-config\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680163 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-console-config\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680193 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9feab1ba-d666-44c5-b929-83a1732c8014-machine-approver-tls\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680212 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-client-ca\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680230 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-config\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680248 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjv52\" (UniqueName: \"kubernetes.io/projected/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-kube-api-access-hjv52\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680267 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-etcd-client\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680287 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-oauth-config\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680311 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680334 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4612eb92-6c6b-457b-bab4-da111cac6428-config\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680356 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4612eb92-6c6b-457b-bab4-da111cac6428-trusted-ca\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680378 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/366baa67-abf1-485c-9430-1a891c9e58c7-serving-cert\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680403 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsqxn\" (UniqueName: \"kubernetes.io/projected/3c6ce409-7798-42c3-8f33-f7d2272bdc86-kube-api-access-fsqxn\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680427 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-serving-cert\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680464 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680490 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92wxv\" (UniqueName: \"kubernetes.io/projected/5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81-kube-api-access-92wxv\") pod \"cluster-samples-operator-665b6dd947-fhvv2\" (UID: \"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680515 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qr2l\" (UniqueName: \"kubernetes.io/projected/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-kube-api-access-5qr2l\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680771 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-encryption-config\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680885 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680936 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65mn5\" (UniqueName: \"kubernetes.io/projected/cacb4d41-218a-441d-a62a-77338425aa3e-kube-api-access-65mn5\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.680981 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681015 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681043 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-audit-dir\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681074 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feab1ba-d666-44c5-b929-83a1732c8014-config\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681100 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681118 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fhvv2\" (UID: \"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681152 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681174 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681195 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-images\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681236 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681267 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-config\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681287 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-service-ca-bundle\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681303 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-image-import-ca\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681322 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cacb4d41-218a-441d-a62a-77338425aa3e-audit-dir\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681357 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c6ce409-7798-42c3-8f33-f7d2272bdc86-serving-cert\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681382 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-etcd-serving-ca\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681409 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvmjk\" (UniqueName: \"kubernetes.io/projected/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-kube-api-access-bvmjk\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681444 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-etcd-client\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681494 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681514 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-audit-policies\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.681530 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-encryption-config\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.686290 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.694174 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.694474 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.694662 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.694811 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.694975 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.695117 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.695446 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.695530 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.695883 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.696027 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.696149 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.696334 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.695452 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.696341 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.696714 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.696776 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.696903 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697056 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697090 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697103 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697064 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697215 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697237 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697329 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697345 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.697524 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.699446 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rt65v"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.706281 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-rjqzp"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.707112 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.707885 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.709540 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.711913 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rc9fc"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.712108 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.714086 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.714663 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.715091 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.715347 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.719025 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.735951 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.737014 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.737815 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.738706 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.739087 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.740713 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.750190 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.750854 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.751343 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.751782 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.754504 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.756334 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.758523 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.764959 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.765386 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.769887 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.779941 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.780744 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.783966 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5pqc\" (UniqueName: \"kubernetes.io/projected/4612eb92-6c6b-457b-bab4-da111cac6428-kube-api-access-b5pqc\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784085 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5819a852-b138-4647-bfeb-186464e53a17-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784221 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-config\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784280 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-trusted-ca-bundle\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784331 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-client-ca\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784385 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/75b0c6fb-238c-4f52-927e-920b288a5f85-metrics-tls\") pod \"dns-operator-744455d44c-xrzvw\" (UID: \"75b0c6fb-238c-4f52-927e-920b288a5f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784439 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9feab1ba-d666-44c5-b929-83a1732c8014-machine-approver-tls\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784494 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-client-ca\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784521 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-console-config\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784626 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjv52\" (UniqueName: \"kubernetes.io/projected/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-kube-api-access-hjv52\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784686 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c67409be-2144-4b3d-be8c-81aaca215420-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784710 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmtrn\" (UniqueName: \"kubernetes.io/projected/c67409be-2144-4b3d-be8c-81aaca215420-kube-api-access-nmtrn\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784758 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxhfw\" (UniqueName: \"kubernetes.io/projected/b6181566-7e6d-42a7-98cc-9cd735da6504-kube-api-access-kxhfw\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784805 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-etcd-service-ca\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784860 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mnk2\" (UniqueName: \"kubernetes.io/projected/44bee603-4ff5-4319-aa19-d5737abba431-kube-api-access-2mnk2\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784886 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-config\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784909 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784929 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4612eb92-6c6b-457b-bab4-da111cac6428-config\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.784956 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4612eb92-6c6b-457b-bab4-da111cac6428-trusted-ca\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.785000 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/366baa67-abf1-485c-9430-1a891c9e58c7-serving-cert\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.786134 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-config\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.787042 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-client-ca\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.787797 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4612eb92-6c6b-457b-bab4-da111cac6428-trusted-ca\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.787847 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-console-config\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.790174 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4612eb92-6c6b-457b-bab4-da111cac6428-config\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.790340 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-trusted-ca-bundle\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.785023 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-etcd-client\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.791993 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-oauth-config\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792042 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsqxn\" (UniqueName: \"kubernetes.io/projected/3c6ce409-7798-42c3-8f33-f7d2272bdc86-kube-api-access-fsqxn\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792080 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbjvq\" (UniqueName: \"kubernetes.io/projected/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-kube-api-access-vbjvq\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792235 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5819a852-b138-4647-bfeb-186464e53a17-metrics-tls\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792274 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792341 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92wxv\" (UniqueName: \"kubernetes.io/projected/5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81-kube-api-access-92wxv\") pod \"cluster-samples-operator-665b6dd947-fhvv2\" (UID: \"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792368 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qr2l\" (UniqueName: \"kubernetes.io/projected/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-kube-api-access-5qr2l\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792399 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-serving-cert\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792454 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44bee603-4ff5-4319-aa19-d5737abba431-serving-cert\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792529 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-encryption-config\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792580 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-config\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792669 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.792965 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65mn5\" (UniqueName: \"kubernetes.io/projected/cacb4d41-218a-441d-a62a-77338425aa3e-kube-api-access-65mn5\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793023 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793077 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793130 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-audit-dir\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793215 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5819a852-b138-4647-bfeb-186464e53a17-trusted-ca\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793250 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feab1ba-d666-44c5-b929-83a1732c8014-config\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793281 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fhvv2\" (UID: \"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793315 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793343 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793435 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-config\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793464 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f00b63a-3d97-443c-9613-e94b0cad6f9b-config\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793580 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793651 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-images\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793714 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6181566-7e6d-42a7-98cc-9cd735da6504-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.794795 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-config\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.796573 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.797586 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.798507 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.798934 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.799141 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-images\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.800262 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801026 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-audit-dir\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.793746 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801227 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/44bee603-4ff5-4319-aa19-d5737abba431-etcd-client\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801329 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-config\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801417 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-service-ca-bundle\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801597 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-image-import-ca\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801702 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801784 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c6ce409-7798-42c3-8f33-f7d2272bdc86-serving-cert\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.801825 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.800710 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-etcd-client\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.803125 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9feab1ba-d666-44c5-b929-83a1732c8014-machine-approver-tls\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.805535 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.811307 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-oauth-config\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.800070 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-serving-cert\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.814197 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fhvv2\" (UID: \"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.814810 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-config\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.814832 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-service-ca-bundle\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.814857 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m2j8j"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.815727 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-image-import-ca\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.815939 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.816701 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.816938 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.816800 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/366baa67-abf1-485c-9430-1a891c9e58c7-serving-cert\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817070 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-etcd-serving-ca\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817158 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cacb4d41-218a-441d-a62a-77338425aa3e-audit-dir\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817242 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817361 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-audit-policies\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817450 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvmjk\" (UniqueName: \"kubernetes.io/projected/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-kube-api-access-bvmjk\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817526 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-etcd-client\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817605 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f00b63a-3d97-443c-9613-e94b0cad6f9b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817705 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b6181566-7e6d-42a7-98cc-9cd735da6504-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817784 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-encryption-config\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817856 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817941 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4612eb92-6c6b-457b-bab4-da111cac6428-serving-cert\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818194 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/748ee850-8101-42cd-b903-4c1443fa32e4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818267 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/748ee850-8101-42cd-b903-4c1443fa32e4-config\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818341 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6181566-7e6d-42a7-98cc-9cd735da6504-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818442 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-dir\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818525 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4wtk\" (UniqueName: \"kubernetes.io/projected/cc0c9cff-8b28-4d9c-b50d-0ab81e16c841-kube-api-access-d4wtk\") pod \"downloads-7954f5f757-pr5cf\" (UID: \"cc0c9cff-8b28-4d9c-b50d-0ab81e16c841\") " pod="openshift-console/downloads-7954f5f757-pr5cf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818592 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818601 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-config\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818769 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cacb4d41-218a-441d-a62a-77338425aa3e-node-pullsecrets\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.819666 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c67409be-2144-4b3d-be8c-81aaca215420-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.819758 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cacb4d41-218a-441d-a62a-77338425aa3e-audit-dir\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.818576 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-etcd-serving-ca\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.819742 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-dir\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.819770 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/748ee850-8101-42cd-b903-4c1443fa32e4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820025 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820052 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-audit-policies\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820063 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-serving-cert\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.817672 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9feab1ba-d666-44c5-b929-83a1732c8014-config\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820088 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820128 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-serving-cert\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820172 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820192 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820218 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820242 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-service-ca\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820325 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6gj9\" (UniqueName: \"kubernetes.io/projected/75b0c6fb-238c-4f52-927e-920b288a5f85-kube-api-access-h6gj9\") pod \"dns-operator-744455d44c-xrzvw\" (UID: \"75b0c6fb-238c-4f52-927e-920b288a5f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820418 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9feab1ba-d666-44c5-b929-83a1732c8014-auth-proxy-config\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820460 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-policies\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820492 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-etcd-ca\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820514 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f00b63a-3d97-443c-9613-e94b0cad6f9b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820535 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbr8s\" (UniqueName: \"kubernetes.io/projected/4d1c19bc-9031-49eb-88a1-3cb263134b3d-kube-api-access-mbr8s\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820554 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-oauth-serving-cert\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820606 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twb6k\" (UniqueName: \"kubernetes.io/projected/9feab1ba-d666-44c5-b929-83a1732c8014-kube-api-access-twb6k\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820626 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820660 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stvcv\" (UniqueName: \"kubernetes.io/projected/5819a852-b138-4647-bfeb-186464e53a17-kube-api-access-stvcv\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820693 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820733 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-serving-cert\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820749 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2198f599-a3a9-472a-b63e-40e302516c64-serving-cert\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820766 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcnx9\" (UniqueName: \"kubernetes.io/projected/2198f599-a3a9-472a-b63e-40e302516c64-kube-api-access-rcnx9\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820785 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820804 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4796\" (UniqueName: \"kubernetes.io/projected/366baa67-abf1-485c-9430-1a891c9e58c7-kube-api-access-s4796\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820822 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3c6ce409-7798-42c3-8f33-f7d2272bdc86-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820844 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrzqz\" (UniqueName: \"kubernetes.io/projected/0b433b98-d626-4286-a89b-3618ba63b58a-kube-api-access-nrzqz\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.820867 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-audit\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.821238 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-service-ca\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.821310 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.821627 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.821702 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-audit\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.821735 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.821837 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cacb4d41-218a-441d-a62a-77338425aa3e-config\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.822249 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-policies\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.822529 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.822616 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2198f599-a3a9-472a-b63e-40e302516c64-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.822874 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9feab1ba-d666-44c5-b929-83a1732c8014-auth-proxy-config\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.823443 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-etcd-client\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.823718 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.824230 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cacb4d41-218a-441d-a62a-77338425aa3e-node-pullsecrets\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.819689 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-encryption-config\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.824432 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.824959 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c6ce409-7798-42c3-8f33-f7d2272bdc86-serving-cert\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.825064 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-serving-cert\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.824657 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.825114 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.825661 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-oauth-serving-cert\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.825765 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3c6ce409-7798-42c3-8f33-f7d2272bdc86-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.825946 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-encryption-config\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.826466 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4612eb92-6c6b-457b-bab4-da111cac6428-serving-cert\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.827250 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cwrvv"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.827926 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cacb4d41-218a-441d-a62a-77338425aa3e-serving-cert\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.828860 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.829998 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.829816 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.829894 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2198f599-a3a9-472a-b63e-40e302516c64-serving-cert\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.830277 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.831601 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.831685 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-sdqq6"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.832188 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.832864 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.833819 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.834068 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jzchz"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.832670 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.834688 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.836024 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.837032 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.837598 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.838014 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.838508 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.839106 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.841477 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.842017 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.843840 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sdrxk"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.844041 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-srrlc"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.845096 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.846078 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.847326 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5b9t"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.848293 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.849887 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.851148 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.852106 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.853389 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.854582 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.855920 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.857026 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xrzvw"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.858246 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-plmm7"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.859563 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-sdqq6"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.860688 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.862224 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.863332 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jfck7"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.864617 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.866315 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vg9l6"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.867903 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.869915 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.870015 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6v99v"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.870812 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6v99v" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.871694 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-5l26w"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.872236 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.873862 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cwrvv"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.876683 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.879681 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.881471 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.882710 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.884003 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jzchz"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.885940 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.897254 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.900941 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5l26w"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.902144 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m2j8j"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.903315 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.904496 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5b9t"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.905624 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6v99v"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.906881 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-mrp4g"] Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.907862 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.915002 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921587 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5819a852-b138-4647-bfeb-186464e53a17-trusted-ca\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921653 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-config\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921683 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f00b63a-3d97-443c-9613-e94b0cad6f9b-config\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921746 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6181566-7e6d-42a7-98cc-9cd735da6504-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921775 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/44bee603-4ff5-4319-aa19-d5737abba431-etcd-client\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921812 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f00b63a-3d97-443c-9613-e94b0cad6f9b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921836 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921860 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b6181566-7e6d-42a7-98cc-9cd735da6504-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921903 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/748ee850-8101-42cd-b903-4c1443fa32e4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921929 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/748ee850-8101-42cd-b903-4c1443fa32e4-config\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921956 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6181566-7e6d-42a7-98cc-9cd735da6504-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.921996 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c67409be-2144-4b3d-be8c-81aaca215420-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922024 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/748ee850-8101-42cd-b903-4c1443fa32e4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922070 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-serving-cert\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922118 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6gj9\" (UniqueName: \"kubernetes.io/projected/75b0c6fb-238c-4f52-927e-920b288a5f85-kube-api-access-h6gj9\") pod \"dns-operator-744455d44c-xrzvw\" (UID: \"75b0c6fb-238c-4f52-927e-920b288a5f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922154 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-etcd-ca\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922181 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f00b63a-3d97-443c-9613-e94b0cad6f9b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922219 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stvcv\" (UniqueName: \"kubernetes.io/projected/5819a852-b138-4647-bfeb-186464e53a17-kube-api-access-stvcv\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922285 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5819a852-b138-4647-bfeb-186464e53a17-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922342 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-client-ca\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922371 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/75b0c6fb-238c-4f52-927e-920b288a5f85-metrics-tls\") pod \"dns-operator-744455d44c-xrzvw\" (UID: \"75b0c6fb-238c-4f52-927e-920b288a5f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922458 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c67409be-2144-4b3d-be8c-81aaca215420-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922518 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmtrn\" (UniqueName: \"kubernetes.io/projected/c67409be-2144-4b3d-be8c-81aaca215420-kube-api-access-nmtrn\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922558 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxhfw\" (UniqueName: \"kubernetes.io/projected/b6181566-7e6d-42a7-98cc-9cd735da6504-kube-api-access-kxhfw\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922586 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-etcd-service-ca\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922617 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mnk2\" (UniqueName: \"kubernetes.io/projected/44bee603-4ff5-4319-aa19-d5737abba431-kube-api-access-2mnk2\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922693 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbjvq\" (UniqueName: \"kubernetes.io/projected/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-kube-api-access-vbjvq\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922720 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5819a852-b138-4647-bfeb-186464e53a17-metrics-tls\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922775 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44bee603-4ff5-4319-aa19-d5737abba431-serving-cert\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922803 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-config\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.922884 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f00b63a-3d97-443c-9613-e94b0cad6f9b-config\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.923145 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-config\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.923511 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6181566-7e6d-42a7-98cc-9cd735da6504-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.923757 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-config\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.924467 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-etcd-service-ca\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.925292 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-client-ca\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.926082 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/44bee603-4ff5-4319-aa19-d5737abba431-etcd-client\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.926394 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.926569 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/44bee603-4ff5-4319-aa19-d5737abba431-etcd-ca\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.927264 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c67409be-2144-4b3d-be8c-81aaca215420-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.927599 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f00b63a-3d97-443c-9613-e94b0cad6f9b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.928111 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c67409be-2144-4b3d-be8c-81aaca215420-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.928126 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-serving-cert\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.928555 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44bee603-4ff5-4319-aa19-d5737abba431-serving-cert\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.932192 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.950892 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.972677 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 19 08:45:58 crc kubenswrapper[4675]: I0219 08:45:58.989970 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.014456 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.027881 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6181566-7e6d-42a7-98cc-9cd735da6504-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.029854 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.050587 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.070884 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.077716 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/75b0c6fb-238c-4f52-927e-920b288a5f85-metrics-tls\") pod \"dns-operator-744455d44c-xrzvw\" (UID: \"75b0c6fb-238c-4f52-927e-920b288a5f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.090368 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.110097 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.130280 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.150147 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.158060 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5819a852-b138-4647-bfeb-186464e53a17-metrics-tls\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.170572 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.196449 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.204834 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5819a852-b138-4647-bfeb-186464e53a17-trusted-ca\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.210485 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.230629 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.250288 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.257199 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/748ee850-8101-42cd-b903-4c1443fa32e4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.269807 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.275046 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/748ee850-8101-42cd-b903-4c1443fa32e4-config\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.311103 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.330969 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.350927 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.370354 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.390463 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.410579 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.430053 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.450724 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.470131 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.490431 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.510315 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.530408 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.550197 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.570201 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.590180 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.609614 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.630381 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.649267 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.670568 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.690453 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.724188 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5pqc\" (UniqueName: \"kubernetes.io/projected/4612eb92-6c6b-457b-bab4-da111cac6428-kube-api-access-b5pqc\") pod \"console-operator-58897d9998-6nnjw\" (UID: \"4612eb92-6c6b-457b-bab4-da111cac6428\") " pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.743236 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjv52\" (UniqueName: \"kubernetes.io/projected/7c39c460-3de5-4b4a-9c38-76ad1de9cbce-kube-api-access-hjv52\") pod \"machine-api-operator-5694c8668f-q6tbj\" (UID: \"7c39c460-3de5-4b4a-9c38-76ad1de9cbce\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.764054 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsqxn\" (UniqueName: \"kubernetes.io/projected/3c6ce409-7798-42c3-8f33-f7d2272bdc86-kube-api-access-fsqxn\") pod \"openshift-config-operator-7777fb866f-rt65v\" (UID: \"3c6ce409-7798-42c3-8f33-f7d2272bdc86\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.770274 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.790430 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.808462 4675 request.go:700] Waited for 1.012303631s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.816291 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92wxv\" (UniqueName: \"kubernetes.io/projected/5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81-kube-api-access-92wxv\") pod \"cluster-samples-operator-665b6dd947-fhvv2\" (UID: \"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.830036 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qr2l\" (UniqueName: \"kubernetes.io/projected/e1ca2913-e5e0-4968-b32e-593e4f8b8fcd-kube-api-access-5qr2l\") pod \"openshift-apiserver-operator-796bbdcf4f-dgvc8\" (UID: \"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.830437 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.850933 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.851557 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.871394 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.890962 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.911195 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.931728 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.932815 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.940838 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.952147 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.955493 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" Feb 19 08:45:59 crc kubenswrapper[4675]: I0219 08:45:59.971425 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.009936 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.012861 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65mn5\" (UniqueName: \"kubernetes.io/projected/cacb4d41-218a-441d-a62a-77338425aa3e-kube-api-access-65mn5\") pod \"apiserver-76f77b778f-sdrxk\" (UID: \"cacb4d41-218a-441d-a62a-77338425aa3e\") " pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.015956 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6nnjw"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.031012 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.080195 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6nnjw" event={"ID":"4612eb92-6c6b-457b-bab4-da111cac6428","Type":"ContainerStarted","Data":"78eb371f227d971f118b811601c8b218f3fb1175e471f8bad240c559b7b08cd9"} Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.080319 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4wtk\" (UniqueName: \"kubernetes.io/projected/cc0c9cff-8b28-4d9c-b50d-0ab81e16c841-kube-api-access-d4wtk\") pod \"downloads-7954f5f757-pr5cf\" (UID: \"cc0c9cff-8b28-4d9c-b50d-0ab81e16c841\") " pod="openshift-console/downloads-7954f5f757-pr5cf" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.084572 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rt65v"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.089812 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvmjk\" (UniqueName: \"kubernetes.io/projected/18324d8e-bab7-4c65-9ad8-32878ffd7b0b-kube-api-access-bvmjk\") pod \"apiserver-7bbb656c7d-99sr2\" (UID: \"18324d8e-bab7-4c65-9ad8-32878ffd7b0b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.107997 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcnx9\" (UniqueName: \"kubernetes.io/projected/2198f599-a3a9-472a-b63e-40e302516c64-kube-api-access-rcnx9\") pod \"authentication-operator-69f744f599-hhkkj\" (UID: \"2198f599-a3a9-472a-b63e-40e302516c64\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.126322 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-pr5cf" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.126679 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrzqz\" (UniqueName: \"kubernetes.io/projected/0b433b98-d626-4286-a89b-3618ba63b58a-kube-api-access-nrzqz\") pod \"console-f9d7485db-vg9l6\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:46:00 crc kubenswrapper[4675]: W0219 08:46:00.127604 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c6ce409_7798_42c3_8f33_f7d2272bdc86.slice/crio-a552881bb73226fcdb80728033c92504e38e1acca5c612b6bcf259a3ccf77dde WatchSource:0}: Error finding container a552881bb73226fcdb80728033c92504e38e1acca5c612b6bcf259a3ccf77dde: Status 404 returned error can't find the container with id a552881bb73226fcdb80728033c92504e38e1acca5c612b6bcf259a3ccf77dde Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.140063 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.145381 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4796\" (UniqueName: \"kubernetes.io/projected/366baa67-abf1-485c-9430-1a891c9e58c7-kube-api-access-s4796\") pod \"route-controller-manager-6576b87f9c-rrj8q\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.176198 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbr8s\" (UniqueName: \"kubernetes.io/projected/4d1c19bc-9031-49eb-88a1-3cb263134b3d-kube-api-access-mbr8s\") pod \"oauth-openshift-558db77b4-plmm7\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.196531 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.199602 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twb6k\" (UniqueName: \"kubernetes.io/projected/9feab1ba-d666-44c5-b929-83a1732c8014-kube-api-access-twb6k\") pod \"machine-approver-56656f9798-rpfbt\" (UID: \"9feab1ba-d666-44c5-b929-83a1732c8014\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.199763 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.207053 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.211291 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.234163 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.254767 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.271177 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.271182 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.282621 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.292559 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.309935 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.316026 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-pr5cf"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.331196 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.349993 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.358672 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.363102 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hhkkj"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.372421 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.390739 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.410079 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q6tbj"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.410762 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.411080 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.415318 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.430911 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.436359 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2"] Feb 19 08:46:00 crc kubenswrapper[4675]: W0219 08:46:00.440966 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c39c460_3de5_4b4a_9c38_76ad1de9cbce.slice/crio-0d7ef28f957eefc2164092e5e69d048de1794e3ffcf9a54f11e04e1c1beadab2 WatchSource:0}: Error finding container 0d7ef28f957eefc2164092e5e69d048de1794e3ffcf9a54f11e04e1c1beadab2: Status 404 returned error can't find the container with id 0d7ef28f957eefc2164092e5e69d048de1794e3ffcf9a54f11e04e1c1beadab2 Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.450835 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.466840 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.479017 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.491510 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.512223 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.530995 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.550603 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.569950 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.590603 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.612211 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.630873 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.649838 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.655149 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q"] Feb 19 08:46:00 crc kubenswrapper[4675]: W0219 08:46:00.662901 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod366baa67_abf1_485c_9430_1a891c9e58c7.slice/crio-54113fdc183c119482cb7c7bc4855a3bdecaee5ccaf07fbc9d78ffcc5feaee3a WatchSource:0}: Error finding container 54113fdc183c119482cb7c7bc4855a3bdecaee5ccaf07fbc9d78ffcc5feaee3a: Status 404 returned error can't find the container with id 54113fdc183c119482cb7c7bc4855a3bdecaee5ccaf07fbc9d78ffcc5feaee3a Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.669880 4675 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.690840 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.712749 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.730485 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.741287 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sdrxk"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.742495 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-plmm7"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.750422 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.773681 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.784258 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vg9l6"] Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.792049 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.809390 4675 request.go:700] Waited for 1.936784576s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.811514 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 19 08:46:00 crc kubenswrapper[4675]: W0219 08:46:00.835969 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b433b98_d626_4286_a89b_3618ba63b58a.slice/crio-d89451d205cba4e1932026342cd8c433ec532c8d5a6745f1891cf489060e0a16 WatchSource:0}: Error finding container d89451d205cba4e1932026342cd8c433ec532c8d5a6745f1891cf489060e0a16: Status 404 returned error can't find the container with id d89451d205cba4e1932026342cd8c433ec532c8d5a6745f1891cf489060e0a16 Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.850566 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.871136 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.891308 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.929556 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f00b63a-3d97-443c-9613-e94b0cad6f9b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tf9mf\" (UID: \"3f00b63a-3d97-443c-9613-e94b0cad6f9b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.947685 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b6181566-7e6d-42a7-98cc-9cd735da6504-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.965869 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxhfw\" (UniqueName: \"kubernetes.io/projected/b6181566-7e6d-42a7-98cc-9cd735da6504-kube-api-access-kxhfw\") pod \"cluster-image-registry-operator-dc59b4c8b-ntvd2\" (UID: \"b6181566-7e6d-42a7-98cc-9cd735da6504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.975361 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" Feb 19 08:46:00 crc kubenswrapper[4675]: I0219 08:46:00.984926 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmtrn\" (UniqueName: \"kubernetes.io/projected/c67409be-2144-4b3d-be8c-81aaca215420-kube-api-access-nmtrn\") pod \"openshift-controller-manager-operator-756b6f6bc6-dj8qx\" (UID: \"c67409be-2144-4b3d-be8c-81aaca215420\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:00.994848 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.006419 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stvcv\" (UniqueName: \"kubernetes.io/projected/5819a852-b138-4647-bfeb-186464e53a17-kube-api-access-stvcv\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.028331 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5819a852-b138-4647-bfeb-186464e53a17-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b9t9z\" (UID: \"5819a852-b138-4647-bfeb-186464e53a17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.051298 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mnk2\" (UniqueName: \"kubernetes.io/projected/44bee603-4ff5-4319-aa19-d5737abba431-kube-api-access-2mnk2\") pod \"etcd-operator-b45778765-rc9fc\" (UID: \"44bee603-4ff5-4319-aa19-d5737abba431\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.067550 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbjvq\" (UniqueName: \"kubernetes.io/projected/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-kube-api-access-vbjvq\") pod \"controller-manager-879f6c89f-srrlc\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.088275 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6gj9\" (UniqueName: \"kubernetes.io/projected/75b0c6fb-238c-4f52-927e-920b288a5f85-kube-api-access-h6gj9\") pod \"dns-operator-744455d44c-xrzvw\" (UID: \"75b0c6fb-238c-4f52-927e-920b288a5f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.098581 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-pr5cf" event={"ID":"cc0c9cff-8b28-4d9c-b50d-0ab81e16c841","Type":"ContainerStarted","Data":"01dd217d7ba785bfbc0138b9ccdad5ea2099b1c193051b08f13f615ba1505d21"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.098670 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-pr5cf" event={"ID":"cc0c9cff-8b28-4d9c-b50d-0ab81e16c841","Type":"ContainerStarted","Data":"d6da214b8d47ff3eab2dea48e682cdfbe3bc051e24ebaac10fda927efdfcf917"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.098831 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-pr5cf" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.101570 4675 patch_prober.go:28] interesting pod/downloads-7954f5f757-pr5cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.101632 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pr5cf" podUID="cc0c9cff-8b28-4d9c-b50d-0ab81e16c841" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.113560 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/748ee850-8101-42cd-b903-4c1443fa32e4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-b8w5h\" (UID: \"748ee850-8101-42cd-b903-4c1443fa32e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.117928 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" event={"ID":"9feab1ba-d666-44c5-b929-83a1732c8014","Type":"ContainerStarted","Data":"107242022b8ccfe0272e489e4a950dc1c304a55959887557c12f491a4483c654"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.117975 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" event={"ID":"9feab1ba-d666-44c5-b929-83a1732c8014","Type":"ContainerStarted","Data":"a7359e758b96bc9e2b3665cd69bc6651e43039c2f29faba3e3f07af8e0418f01"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.121619 4675 generic.go:334] "Generic (PLEG): container finished" podID="18324d8e-bab7-4c65-9ad8-32878ffd7b0b" containerID="3be8d58d6c05bf347003825c2667517b8d6d6f3e60dc4fbd12dedeac9d14eba1" exitCode=0 Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.121689 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" event={"ID":"18324d8e-bab7-4c65-9ad8-32878ffd7b0b","Type":"ContainerDied","Data":"3be8d58d6c05bf347003825c2667517b8d6d6f3e60dc4fbd12dedeac9d14eba1"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.121711 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" event={"ID":"18324d8e-bab7-4c65-9ad8-32878ffd7b0b","Type":"ContainerStarted","Data":"c24032009008c6bd6b0ffd70e9af7d702c865820470d0ceaceabf35c93213920"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.139095 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" event={"ID":"7c39c460-3de5-4b4a-9c38-76ad1de9cbce","Type":"ContainerStarted","Data":"788d061447930b08f783cdc31b79687e61c6ce935b0448d3d51c21b1c9833480"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.139152 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" event={"ID":"7c39c460-3de5-4b4a-9c38-76ad1de9cbce","Type":"ContainerStarted","Data":"903ac97249562550df60111fe153b3710c8ebe35477090652992aec98ca5c6d4"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.139165 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" event={"ID":"7c39c460-3de5-4b4a-9c38-76ad1de9cbce","Type":"ContainerStarted","Data":"0d7ef28f957eefc2164092e5e69d048de1794e3ffcf9a54f11e04e1c1beadab2"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.142897 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" event={"ID":"366baa67-abf1-485c-9430-1a891c9e58c7","Type":"ContainerStarted","Data":"b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.142965 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" event={"ID":"366baa67-abf1-485c-9430-1a891c9e58c7","Type":"ContainerStarted","Data":"54113fdc183c119482cb7c7bc4855a3bdecaee5ccaf07fbc9d78ffcc5feaee3a"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.143523 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.148468 4675 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-rrj8q container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.148504 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" podUID="366baa67-abf1-485c-9430-1a891c9e58c7" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.152333 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" event={"ID":"2198f599-a3a9-472a-b63e-40e302516c64","Type":"ContainerStarted","Data":"9056fc2f5688b679cbb5071bcc6e90711a77948e2b2ea644b24c530314b6675b"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.152381 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" event={"ID":"2198f599-a3a9-472a-b63e-40e302516c64","Type":"ContainerStarted","Data":"6dd9c364876a707f31f9108a6ec3e08c50525ccbf4a5e0148f0bffeec43252f5"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154468 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mzqf\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-kube-api-access-8mzqf\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154548 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-bound-sa-token\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154606 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/08306565-7dde-46be-b2de-d813dfdd976a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154715 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-registry-certificates\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154748 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/08306565-7dde-46be-b2de-d813dfdd976a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154788 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-trusted-ca\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154826 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.154856 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-registry-tls\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.155323 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:01.65530609 +0000 UTC m=+143.282396438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.158211 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" event={"ID":"4d1c19bc-9031-49eb-88a1-3cb263134b3d","Type":"ContainerStarted","Data":"c6a382c9bff1c8c41350d4e8886c3ead1f6718f0a3ac0aa737dba7cae34c8b75"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.168204 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" event={"ID":"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd","Type":"ContainerStarted","Data":"896b13ed80ed713a1fc3273b9215b6c4b7a0a264d8d492eb2027f15a33373bce"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.168273 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" event={"ID":"e1ca2913-e5e0-4968-b32e-593e4f8b8fcd","Type":"ContainerStarted","Data":"58af6637a9332d7cd07ee23be6eee8ef5bfba8568cda06149e5832286a2e6e0b"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.185337 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.191030 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vg9l6" event={"ID":"0b433b98-d626-4286-a89b-3618ba63b58a","Type":"ContainerStarted","Data":"4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.191124 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vg9l6" event={"ID":"0b433b98-d626-4286-a89b-3618ba63b58a","Type":"ContainerStarted","Data":"d89451d205cba4e1932026342cd8c433ec532c8d5a6745f1891cf489060e0a16"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.196353 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6nnjw" event={"ID":"4612eb92-6c6b-457b-bab4-da111cac6428","Type":"ContainerStarted","Data":"db015e89ecf4b4d07b8483190de92ae271114853497b7351be5c491caf892945"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.196399 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.211768 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" event={"ID":"3c6ce409-7798-42c3-8f33-f7d2272bdc86","Type":"ContainerDied","Data":"99ff3624e5ec73b81f962def1836c969fda78224438984b161c09074ceadc8e9"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.212270 4675 generic.go:334] "Generic (PLEG): container finished" podID="3c6ce409-7798-42c3-8f33-f7d2272bdc86" containerID="99ff3624e5ec73b81f962def1836c969fda78224438984b161c09074ceadc8e9" exitCode=0 Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.212412 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" event={"ID":"3c6ce409-7798-42c3-8f33-f7d2272bdc86","Type":"ContainerStarted","Data":"a552881bb73226fcdb80728033c92504e38e1acca5c612b6bcf259a3ccf77dde"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.219343 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" event={"ID":"cacb4d41-218a-441d-a62a-77338425aa3e","Type":"ContainerStarted","Data":"6118e6bc548fe4b0e29c1fb5fd4ce321ea4c7911c7baad61e2de097e603878ea"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.247080 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2"] Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.249335 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" event={"ID":"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81","Type":"ContainerStarted","Data":"e22dcee9620087700310cde3a55dd22d8ae44283b1759396bd609bbd4ce777d6"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.249397 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" event={"ID":"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81","Type":"ContainerStarted","Data":"5ab0d2adb38a7bdcfa59a54c145ad13f79342f1545f008763db168e0140ed8eb"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.249410 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" event={"ID":"5f1d04f4-8d09-43c2-a1a2-dc87f6a08d81","Type":"ContainerStarted","Data":"c8408bbd0718facf13d949454c4b782c3c202fb64f27bb72600762dd184fdbc1"} Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.255589 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256077 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-auth-proxy-config\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256130 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/08306565-7dde-46be-b2de-d813dfdd976a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256153 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f2da5f24-2b80-4967-8a86-ab370a38e45f-tmpfs\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256175 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-mountpoint-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256252 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-socket-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256318 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-secret-volume\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256364 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256386 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/86e4bad1-6678-41f1-bd97-f48fc5312b10-srv-cert\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.256498 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:01.756475318 +0000 UTC m=+143.383565656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256568 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjg4p\" (UniqueName: \"kubernetes.io/projected/230232aa-5a5d-4436-a495-ecd2919fa56b-kube-api-access-sjg4p\") pod \"multus-admission-controller-857f4d67dd-m2j8j\" (UID: \"230232aa-5a5d-4436-a495-ecd2919fa56b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256608 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-registry-tls\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256685 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2mxz\" (UniqueName: \"kubernetes.io/projected/f2da5f24-2b80-4967-8a86-ab370a38e45f-kube-api-access-d2mxz\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256733 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8j6n\" (UniqueName: \"kubernetes.io/projected/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-kube-api-access-r8j6n\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256769 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4gvf\" (UniqueName: \"kubernetes.io/projected/d3d57965-61a8-4a6f-883b-20360fd36509-kube-api-access-v4gvf\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256853 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256951 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mzqf\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-kube-api-access-8mzqf\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.256982 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqnsh\" (UniqueName: \"kubernetes.io/projected/bbed8b9a-ef34-424b-81e0-457a38384ba0-kube-api-access-zqnsh\") pod \"control-plane-machine-set-operator-78cbb6b69f-wf8zz\" (UID: \"bbed8b9a-ef34-424b-81e0-457a38384ba0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257063 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s8zx\" (UniqueName: \"kubernetes.io/projected/24727689-f73c-4ff3-9b0c-9fa625d16c27-kube-api-access-5s8zx\") pod \"package-server-manager-789f6589d5-wt2kk\" (UID: \"24727689-f73c-4ff3-9b0c-9fa625d16c27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257109 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhf8p\" (UniqueName: \"kubernetes.io/projected/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-kube-api-access-hhf8p\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257128 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-default-certificate\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257168 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-csi-data-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257197 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/914e8e63-cb4b-4784-84be-fa509224d2f1-config-volume\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257256 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-config\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257275 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d35a2814-4eb0-4e49-be8c-bbde18573d00-srv-cert\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257351 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-config-volume\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257371 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-plugins-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257434 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257451 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-proxy-tls\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257496 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hn69\" (UniqueName: \"kubernetes.io/projected/914e8e63-cb4b-4784-84be-fa509224d2f1-kube-api-access-4hn69\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257523 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-registry-certificates\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257520 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.258425 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/08306565-7dde-46be-b2de-d813dfdd976a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.257539 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqwsv\" (UniqueName: \"kubernetes.io/projected/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-kube-api-access-bqwsv\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.258889 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbbr\" (UniqueName: \"kubernetes.io/projected/017fb59c-7331-42bf-9a21-8dff74814740-kube-api-access-sgbbr\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.258949 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5acabaa8-91d2-4f21-8a47-3959801a3600-signing-key\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.258968 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-stats-auth\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.258985 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-metrics-certs\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259022 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/816b48eb-9013-414f-a400-909963ff61fe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259069 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6dcz\" (UniqueName: \"kubernetes.io/projected/d4011782-4d60-4178-9ed5-e76a79a672aa-kube-api-access-w6dcz\") pod \"migrator-59844c95c7-t6vv7\" (UID: \"d4011782-4d60-4178-9ed5-e76a79a672aa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259095 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259158 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bbed8b9a-ef34-424b-81e0-457a38384ba0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wf8zz\" (UID: \"bbed8b9a-ef34-424b-81e0-457a38384ba0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259186 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-trusted-ca\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259209 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259229 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-registration-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259325 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/24a4abb3-12f1-4f94-905b-344fc352730c-certs\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259369 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2da5f24-2b80-4967-8a86-ab370a38e45f-webhook-cert\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.259387 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/816b48eb-9013-414f-a400-909963ff61fe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.260716 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-trusted-ca\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.261287 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2da5f24-2b80-4967-8a86-ab370a38e45f-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.261680 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7vlj\" (UniqueName: \"kubernetes.io/projected/5acabaa8-91d2-4f21-8a47-3959801a3600-kube-api-access-f7vlj\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.261746 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n5vh\" (UniqueName: \"kubernetes.io/projected/816b48eb-9013-414f-a400-909963ff61fe-kube-api-access-4n5vh\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.261773 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd8n9\" (UniqueName: \"kubernetes.io/projected/d35a2814-4eb0-4e49-be8c-bbde18573d00-kube-api-access-nd8n9\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.261803 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/24727689-f73c-4ff3-9b0c-9fa625d16c27-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wt2kk\" (UID: \"24727689-f73c-4ff3-9b0c-9fa625d16c27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.261834 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5acabaa8-91d2-4f21-8a47-3959801a3600-signing-cabundle\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.262048 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk2bj\" (UniqueName: \"kubernetes.io/projected/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-kube-api-access-kk2bj\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.262113 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-bound-sa-token\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.262341 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5frts\" (UniqueName: \"kubernetes.io/projected/a80cf4f9-94b9-4694-85e5-9643aace2211-kube-api-access-5frts\") pod \"ingress-canary-5l26w\" (UID: \"a80cf4f9-94b9-4694-85e5-9643aace2211\") " pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.262437 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a80cf4f9-94b9-4694-85e5-9643aace2211-cert\") pod \"ingress-canary-5l26w\" (UID: \"a80cf4f9-94b9-4694-85e5-9643aace2211\") " pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.262480 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rr2c\" (UniqueName: \"kubernetes.io/projected/24a4abb3-12f1-4f94-905b-344fc352730c-kube-api-access-4rr2c\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.262531 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3d57965-61a8-4a6f-883b-20360fd36509-service-ca-bundle\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.262556 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-serving-cert\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: W0219 08:46:01.263194 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6181566_7e6d_42a7_98cc_9cd735da6504.slice/crio-aaae33dd83945105e4fc14e1176946700b9b545abfd484eef299eae126c1ea31 WatchSource:0}: Error finding container aaae33dd83945105e4fc14e1176946700b9b545abfd484eef299eae126c1ea31: Status 404 returned error can't find the container with id aaae33dd83945105e4fc14e1176946700b9b545abfd484eef299eae126c1ea31 Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.270314 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-registry-certificates\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.275135 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.275663 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-registry-tls\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.275832 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/230232aa-5a5d-4436-a495-ecd2919fa56b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m2j8j\" (UID: \"230232aa-5a5d-4436-a495-ecd2919fa56b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.275868 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdgdj\" (UniqueName: \"kubernetes.io/projected/02a5e839-abbb-424e-b702-7e197d9d3307-kube-api-access-sdgdj\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.277562 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-images\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.277628 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/24a4abb3-12f1-4f94-905b-344fc352730c-node-bootstrap-token\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.277694 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w22b\" (UniqueName: \"kubernetes.io/projected/86e4bad1-6678-41f1-bd97-f48fc5312b10-kube-api-access-2w22b\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.277780 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d35a2814-4eb0-4e49-be8c-bbde18573d00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.277810 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/86e4bad1-6678-41f1-bd97-f48fc5312b10-profile-collector-cert\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.277923 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.279110 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/08306565-7dde-46be-b2de-d813dfdd976a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.279206 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/914e8e63-cb4b-4784-84be-fa509224d2f1-metrics-tls\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.279414 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-proxy-tls\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.289620 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/08306565-7dde-46be-b2de-d813dfdd976a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.310440 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf"] Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.313863 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.316488 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mzqf\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-kube-api-access-8mzqf\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.316809 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.321293 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.329070 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-bound-sa-token\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: W0219 08:46:01.366771 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f00b63a_3d97_443c_9613_e94b0cad6f9b.slice/crio-a1b692b8dd9203a6dd694deb34dcca6249552f805fdf819e4fea7aee9334f976 WatchSource:0}: Error finding container a1b692b8dd9203a6dd694deb34dcca6249552f805fdf819e4fea7aee9334f976: Status 404 returned error can't find the container with id a1b692b8dd9203a6dd694deb34dcca6249552f805fdf819e4fea7aee9334f976 Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.369317 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-6nnjw" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381147 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-proxy-tls\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381197 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-auth-proxy-config\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381219 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f2da5f24-2b80-4967-8a86-ab370a38e45f-tmpfs\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381236 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-mountpoint-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381261 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-socket-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381277 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-secret-volume\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381297 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381312 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/86e4bad1-6678-41f1-bd97-f48fc5312b10-srv-cert\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381335 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381354 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjg4p\" (UniqueName: \"kubernetes.io/projected/230232aa-5a5d-4436-a495-ecd2919fa56b-kube-api-access-sjg4p\") pod \"multus-admission-controller-857f4d67dd-m2j8j\" (UID: \"230232aa-5a5d-4436-a495-ecd2919fa56b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381373 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2mxz\" (UniqueName: \"kubernetes.io/projected/f2da5f24-2b80-4967-8a86-ab370a38e45f-kube-api-access-d2mxz\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381390 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8j6n\" (UniqueName: \"kubernetes.io/projected/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-kube-api-access-r8j6n\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381407 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4gvf\" (UniqueName: \"kubernetes.io/projected/d3d57965-61a8-4a6f-883b-20360fd36509-kube-api-access-v4gvf\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381424 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381448 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqnsh\" (UniqueName: \"kubernetes.io/projected/bbed8b9a-ef34-424b-81e0-457a38384ba0-kube-api-access-zqnsh\") pod \"control-plane-machine-set-operator-78cbb6b69f-wf8zz\" (UID: \"bbed8b9a-ef34-424b-81e0-457a38384ba0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381473 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s8zx\" (UniqueName: \"kubernetes.io/projected/24727689-f73c-4ff3-9b0c-9fa625d16c27-kube-api-access-5s8zx\") pod \"package-server-manager-789f6589d5-wt2kk\" (UID: \"24727689-f73c-4ff3-9b0c-9fa625d16c27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381489 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhf8p\" (UniqueName: \"kubernetes.io/projected/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-kube-api-access-hhf8p\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381508 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-default-certificate\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381523 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-csi-data-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381542 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/914e8e63-cb4b-4784-84be-fa509224d2f1-config-volume\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381565 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-config\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381581 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d35a2814-4eb0-4e49-be8c-bbde18573d00-srv-cert\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381601 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-config-volume\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381617 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-plugins-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381645 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381668 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-proxy-tls\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381684 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hn69\" (UniqueName: \"kubernetes.io/projected/914e8e63-cb4b-4784-84be-fa509224d2f1-kube-api-access-4hn69\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381702 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqwsv\" (UniqueName: \"kubernetes.io/projected/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-kube-api-access-bqwsv\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381717 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbbr\" (UniqueName: \"kubernetes.io/projected/017fb59c-7331-42bf-9a21-8dff74814740-kube-api-access-sgbbr\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381734 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5acabaa8-91d2-4f21-8a47-3959801a3600-signing-key\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381752 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-stats-auth\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381769 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-metrics-certs\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381786 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/816b48eb-9013-414f-a400-909963ff61fe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381944 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6dcz\" (UniqueName: \"kubernetes.io/projected/d4011782-4d60-4178-9ed5-e76a79a672aa-kube-api-access-w6dcz\") pod \"migrator-59844c95c7-t6vv7\" (UID: \"d4011782-4d60-4178-9ed5-e76a79a672aa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381965 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.381990 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bbed8b9a-ef34-424b-81e0-457a38384ba0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wf8zz\" (UID: \"bbed8b9a-ef34-424b-81e0-457a38384ba0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382025 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382040 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-registration-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382067 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/24a4abb3-12f1-4f94-905b-344fc352730c-certs\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382083 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2da5f24-2b80-4967-8a86-ab370a38e45f-webhook-cert\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382097 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/816b48eb-9013-414f-a400-909963ff61fe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382116 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2da5f24-2b80-4967-8a86-ab370a38e45f-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382132 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7vlj\" (UniqueName: \"kubernetes.io/projected/5acabaa8-91d2-4f21-8a47-3959801a3600-kube-api-access-f7vlj\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382149 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n5vh\" (UniqueName: \"kubernetes.io/projected/816b48eb-9013-414f-a400-909963ff61fe-kube-api-access-4n5vh\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382164 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd8n9\" (UniqueName: \"kubernetes.io/projected/d35a2814-4eb0-4e49-be8c-bbde18573d00-kube-api-access-nd8n9\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382183 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/24727689-f73c-4ff3-9b0c-9fa625d16c27-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wt2kk\" (UID: \"24727689-f73c-4ff3-9b0c-9fa625d16c27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382199 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5acabaa8-91d2-4f21-8a47-3959801a3600-signing-cabundle\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382217 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk2bj\" (UniqueName: \"kubernetes.io/projected/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-kube-api-access-kk2bj\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382241 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5frts\" (UniqueName: \"kubernetes.io/projected/a80cf4f9-94b9-4694-85e5-9643aace2211-kube-api-access-5frts\") pod \"ingress-canary-5l26w\" (UID: \"a80cf4f9-94b9-4694-85e5-9643aace2211\") " pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382256 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a80cf4f9-94b9-4694-85e5-9643aace2211-cert\") pod \"ingress-canary-5l26w\" (UID: \"a80cf4f9-94b9-4694-85e5-9643aace2211\") " pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382270 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rr2c\" (UniqueName: \"kubernetes.io/projected/24a4abb3-12f1-4f94-905b-344fc352730c-kube-api-access-4rr2c\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382284 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3d57965-61a8-4a6f-883b-20360fd36509-service-ca-bundle\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382300 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-serving-cert\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382317 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/230232aa-5a5d-4436-a495-ecd2919fa56b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m2j8j\" (UID: \"230232aa-5a5d-4436-a495-ecd2919fa56b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382332 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdgdj\" (UniqueName: \"kubernetes.io/projected/02a5e839-abbb-424e-b702-7e197d9d3307-kube-api-access-sdgdj\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382350 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-images\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382365 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/24a4abb3-12f1-4f94-905b-344fc352730c-node-bootstrap-token\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382447 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w22b\" (UniqueName: \"kubernetes.io/projected/86e4bad1-6678-41f1-bd97-f48fc5312b10-kube-api-access-2w22b\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382468 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d35a2814-4eb0-4e49-be8c-bbde18573d00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382483 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/86e4bad1-6678-41f1-bd97-f48fc5312b10-profile-collector-cert\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382510 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.382529 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/914e8e63-cb4b-4784-84be-fa509224d2f1-metrics-tls\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.392591 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:01.892572218 +0000 UTC m=+143.519662486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.392962 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-plugins-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.394923 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-config-volume\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.395382 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-mountpoint-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.397064 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-config\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.398408 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/86e4bad1-6678-41f1-bd97-f48fc5312b10-srv-cert\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.398567 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.398621 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-socket-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.398569 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-csi-data-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.398701 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/914e8e63-cb4b-4784-84be-fa509224d2f1-config-volume\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.399037 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f2da5f24-2b80-4967-8a86-ab370a38e45f-tmpfs\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.401625 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-stats-auth\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.405092 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-auth-proxy-config\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.408510 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.408607 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3d57965-61a8-4a6f-883b-20360fd36509-service-ca-bundle\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.410022 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5acabaa8-91d2-4f21-8a47-3959801a3600-signing-cabundle\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.411189 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/02a5e839-abbb-424e-b702-7e197d9d3307-registration-dir\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.411264 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-proxy-tls\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.412798 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.413849 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/914e8e63-cb4b-4784-84be-fa509224d2f1-metrics-tls\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.414486 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-images\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.421469 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-default-certificate\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.421650 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/816b48eb-9013-414f-a400-909963ff61fe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.424280 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2da5f24-2b80-4967-8a86-ab370a38e45f-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.425054 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d35a2814-4eb0-4e49-be8c-bbde18573d00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.425547 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a80cf4f9-94b9-4694-85e5-9643aace2211-cert\") pod \"ingress-canary-5l26w\" (UID: \"a80cf4f9-94b9-4694-85e5-9643aace2211\") " pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.435372 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/816b48eb-9013-414f-a400-909963ff61fe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.438520 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/24727689-f73c-4ff3-9b0c-9fa625d16c27-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wt2kk\" (UID: \"24727689-f73c-4ff3-9b0c-9fa625d16c27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.450956 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d35a2814-4eb0-4e49-be8c-bbde18573d00-srv-cert\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.450985 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bbed8b9a-ef34-424b-81e0-457a38384ba0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wf8zz\" (UID: \"bbed8b9a-ef34-424b-81e0-457a38384ba0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.451436 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/230232aa-5a5d-4436-a495-ecd2919fa56b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m2j8j\" (UID: \"230232aa-5a5d-4436-a495-ecd2919fa56b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.451557 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3d57965-61a8-4a6f-883b-20360fd36509-metrics-certs\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.451853 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-proxy-tls\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.452152 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.452296 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2da5f24-2b80-4967-8a86-ab370a38e45f-webhook-cert\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.452380 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/24a4abb3-12f1-4f94-905b-344fc352730c-node-bootstrap-token\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.452471 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.452616 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-secret-volume\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.461211 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5acabaa8-91d2-4f21-8a47-3959801a3600-signing-key\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.462254 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/24a4abb3-12f1-4f94-905b-344fc352730c-certs\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.463265 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/86e4bad1-6678-41f1-bd97-f48fc5312b10-profile-collector-cert\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.466712 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-serving-cert\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.467407 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqnsh\" (UniqueName: \"kubernetes.io/projected/bbed8b9a-ef34-424b-81e0-457a38384ba0-kube-api-access-zqnsh\") pod \"control-plane-machine-set-operator-78cbb6b69f-wf8zz\" (UID: \"bbed8b9a-ef34-424b-81e0-457a38384ba0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.481543 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhf8p\" (UniqueName: \"kubernetes.io/projected/f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f-kube-api-access-hhf8p\") pod \"machine-config-controller-84d6567774-s8cwp\" (UID: \"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.481684 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s8zx\" (UniqueName: \"kubernetes.io/projected/24727689-f73c-4ff3-9b0c-9fa625d16c27-kube-api-access-5s8zx\") pod \"package-server-manager-789f6589d5-wt2kk\" (UID: \"24727689-f73c-4ff3-9b0c-9fa625d16c27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.483950 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.484540 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:01.984518037 +0000 UTC m=+143.611608305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.494676 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2mxz\" (UniqueName: \"kubernetes.io/projected/f2da5f24-2b80-4967-8a86-ab370a38e45f-kube-api-access-d2mxz\") pod \"packageserver-d55dfcdfc-d9r67\" (UID: \"f2da5f24-2b80-4967-8a86-ab370a38e45f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.511504 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjg4p\" (UniqueName: \"kubernetes.io/projected/230232aa-5a5d-4436-a495-ecd2919fa56b-kube-api-access-sjg4p\") pod \"multus-admission-controller-857f4d67dd-m2j8j\" (UID: \"230232aa-5a5d-4436-a495-ecd2919fa56b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.529573 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hn69\" (UniqueName: \"kubernetes.io/projected/914e8e63-cb4b-4784-84be-fa509224d2f1-kube-api-access-4hn69\") pod \"dns-default-6v99v\" (UID: \"914e8e63-cb4b-4784-84be-fa509224d2f1\") " pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.567390 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31ea7bfc-f519-411d-ad96-ab0edbbdac0f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vmq9z\" (UID: \"31ea7bfc-f519-411d-ad96-ab0edbbdac0f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.597718 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.598158 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.098139609 +0000 UTC m=+143.725229877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.599796 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8j6n\" (UniqueName: \"kubernetes.io/projected/a7fb92b9-e97b-46a0-b3da-c9186bd4b585-kube-api-access-r8j6n\") pod \"service-ca-operator-777779d784-jzchz\" (UID: \"a7fb92b9-e97b-46a0-b3da-c9186bd4b585\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.606433 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4gvf\" (UniqueName: \"kubernetes.io/projected/d3d57965-61a8-4a6f-883b-20360fd36509-kube-api-access-v4gvf\") pod \"router-default-5444994796-rjqzp\" (UID: \"d3d57965-61a8-4a6f-883b-20360fd36509\") " pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.621297 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbbr\" (UniqueName: \"kubernetes.io/projected/017fb59c-7331-42bf-9a21-8dff74814740-kube-api-access-sgbbr\") pod \"marketplace-operator-79b997595-cwrvv\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.629010 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqwsv\" (UniqueName: \"kubernetes.io/projected/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-kube-api-access-bqwsv\") pod \"collect-profiles-29524845-v9r5n\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.638667 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.649233 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk2bj\" (UniqueName: \"kubernetes.io/projected/f0cbf82a-8336-4b7f-a96e-0329a0dffb40-kube-api-access-kk2bj\") pod \"machine-config-operator-74547568cd-86dwp\" (UID: \"f0cbf82a-8336-4b7f-a96e-0329a0dffb40\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.657916 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-srrlc"] Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.665969 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.679174 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.685887 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.694452 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rr2c\" (UniqueName: \"kubernetes.io/projected/24a4abb3-12f1-4f94-905b-344fc352730c-kube-api-access-4rr2c\") pod \"machine-config-server-mrp4g\" (UID: \"24a4abb3-12f1-4f94-905b-344fc352730c\") " pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.703140 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.703673 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.203625167 +0000 UTC m=+143.830715435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.705096 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.714235 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.714255 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6dcz\" (UniqueName: \"kubernetes.io/projected/d4011782-4d60-4178-9ed5-e76a79a672aa-kube-api-access-w6dcz\") pod \"migrator-59844c95c7-t6vv7\" (UID: \"d4011782-4d60-4178-9ed5-e76a79a672aa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.722299 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.735041 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdgdj\" (UniqueName: \"kubernetes.io/projected/02a5e839-abbb-424e-b702-7e197d9d3307-kube-api-access-sdgdj\") pod \"csi-hostpathplugin-n5b9t\" (UID: \"02a5e839-abbb-424e-b702-7e197d9d3307\") " pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.735348 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.735561 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5frts\" (UniqueName: \"kubernetes.io/projected/a80cf4f9-94b9-4694-85e5-9643aace2211-kube-api-access-5frts\") pod \"ingress-canary-5l26w\" (UID: \"a80cf4f9-94b9-4694-85e5-9643aace2211\") " pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.745739 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.748855 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.756620 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rc9fc"] Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.765907 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.772388 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w22b\" (UniqueName: \"kubernetes.io/projected/86e4bad1-6678-41f1-bd97-f48fc5312b10-kube-api-access-2w22b\") pod \"catalog-operator-68c6474976-p8bgp\" (UID: \"86e4bad1-6678-41f1-bd97-f48fc5312b10\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.786044 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.786402 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n5vh\" (UniqueName: \"kubernetes.io/projected/816b48eb-9013-414f-a400-909963ff61fe-kube-api-access-4n5vh\") pod \"kube-storage-version-migrator-operator-b67b599dd-rj8cg\" (UID: \"816b48eb-9013-414f-a400-909963ff61fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.796443 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.804839 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5l26w" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.805356 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.805695 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.305679054 +0000 UTC m=+143.932769322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.809802 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mrp4g" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.825854 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7vlj\" (UniqueName: \"kubernetes.io/projected/5acabaa8-91d2-4f21-8a47-3959801a3600-kube-api-access-f7vlj\") pod \"service-ca-9c57cc56f-sdqq6\" (UID: \"5acabaa8-91d2-4f21-8a47-3959801a3600\") " pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.826382 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd8n9\" (UniqueName: \"kubernetes.io/projected/d35a2814-4eb0-4e49-be8c-bbde18573d00-kube-api-access-nd8n9\") pod \"olm-operator-6b444d44fb-wtr45\" (UID: \"d35a2814-4eb0-4e49-be8c-bbde18573d00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.840350 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx"] Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.906656 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h"] Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.908390 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:01 crc kubenswrapper[4675]: E0219 08:46:01.908818 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.408781415 +0000 UTC m=+144.035871683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.931508 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xrzvw"] Feb 19 08:46:01 crc kubenswrapper[4675]: W0219 08:46:01.937759 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc67409be_2144_4b3d_be8c_81aaca215420.slice/crio-41d6e868599f1469012af809201ca11380d5b5f8a211b7f5853db7e4834b651a WatchSource:0}: Error finding container 41d6e868599f1469012af809201ca11380d5b5f8a211b7f5853db7e4834b651a: Status 404 returned error can't find the container with id 41d6e868599f1469012af809201ca11380d5b5f8a211b7f5853db7e4834b651a Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.949577 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.957083 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" Feb 19 08:46:01 crc kubenswrapper[4675]: I0219 08:46:01.978600 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.011048 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.011442 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.51140967 +0000 UTC m=+144.138499938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: W0219 08:46:02.023484 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75b0c6fb_238c_4f52_927e_920b288a5f85.slice/crio-af871e29a940a5bb0ab3a9d934ca65e1cf739051718f521d8cfbce475e2d0a55 WatchSource:0}: Error finding container af871e29a940a5bb0ab3a9d934ca65e1cf739051718f521d8cfbce475e2d0a55: Status 404 returned error can't find the container with id af871e29a940a5bb0ab3a9d934ca65e1cf739051718f521d8cfbce475e2d0a55 Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.025222 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.055897 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.069139 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.081786 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.112551 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.112676 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.612655371 +0000 UTC m=+144.239745629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.113244 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.113723 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.613705775 +0000 UTC m=+144.240796043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.176485 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.189299 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.201693 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-pr5cf" podStartSLOduration=123.201665775 podStartE2EDuration="2m3.201665775s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:02.19996897 +0000 UTC m=+143.827059248" watchObservedRunningTime="2026-02-19 08:46:02.201665775 +0000 UTC m=+143.828756043" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.214273 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.214757 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.714742729 +0000 UTC m=+144.341832997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: W0219 08:46:02.220080 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5819a852_b138_4647_bfeb_186464e53a17.slice/crio-9abd4551d908eb6729bc6a7b44b85f5af1c83d745beb2df204a520283319a7f0 WatchSource:0}: Error finding container 9abd4551d908eb6729bc6a7b44b85f5af1c83d745beb2df204a520283319a7f0: Status 404 returned error can't find the container with id 9abd4551d908eb6729bc6a7b44b85f5af1c83d745beb2df204a520283319a7f0 Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.278665 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" event={"ID":"ad9383f1-0293-4f42-b7f9-2e3fef0e4569","Type":"ContainerStarted","Data":"3922889b5e9e5ebe6392c49323b691eec8a8ede0ae2457bfd3d48c164ab159d5"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.287770 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" event={"ID":"75b0c6fb-238c-4f52-927e-920b288a5f85","Type":"ContainerStarted","Data":"af871e29a940a5bb0ab3a9d934ca65e1cf739051718f521d8cfbce475e2d0a55"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.289592 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6tbj" podStartSLOduration=123.289576283 podStartE2EDuration="2m3.289576283s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:02.288029373 +0000 UTC m=+143.915119651" watchObservedRunningTime="2026-02-19 08:46:02.289576283 +0000 UTC m=+143.916666561" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.314538 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" event={"ID":"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f","Type":"ContainerStarted","Data":"f66a1c7340a10e6047c7cc385253e7d3271c5758efadf8a1df6f92330aa50276"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.321037 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.321447 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.821428665 +0000 UTC m=+144.448518933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.337408 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgvc8" podStartSLOduration=123.337384472 podStartE2EDuration="2m3.337384472s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:02.337038741 +0000 UTC m=+143.964129009" watchObservedRunningTime="2026-02-19 08:46:02.337384472 +0000 UTC m=+143.964474740" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.338953 4675 generic.go:334] "Generic (PLEG): container finished" podID="cacb4d41-218a-441d-a62a-77338425aa3e" containerID="1b98913148c4f7003204130fca679211c0d5b4138dd457ce10920f2583ef6074" exitCode=0 Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.339117 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" event={"ID":"cacb4d41-218a-441d-a62a-77338425aa3e","Type":"ContainerDied","Data":"1b98913148c4f7003204130fca679211c0d5b4138dd457ce10920f2583ef6074"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.353850 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-rjqzp" event={"ID":"d3d57965-61a8-4a6f-883b-20360fd36509","Type":"ContainerStarted","Data":"fe9b24e06c998b5f08e369d10efe7f99e03cb65ce071ec637b80ec4cc38e3bd8"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.370535 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" event={"ID":"9feab1ba-d666-44c5-b929-83a1732c8014","Type":"ContainerStarted","Data":"4c1da7253894ed05d946d84ab9ad395b0aca5a98522f9b6b5645c74115bd6b4e"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.374794 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" event={"ID":"f0cbf82a-8336-4b7f-a96e-0329a0dffb40","Type":"ContainerStarted","Data":"2fe156aac9c91e5504a12a5b250eedaf5edfea58b8820f3a13bbb996bfb927d9"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.378367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" event={"ID":"3c6ce409-7798-42c3-8f33-f7d2272bdc86","Type":"ContainerStarted","Data":"517ba2e60b70c98690f2763f5547376502e6fcc0704dcf4b3c8f21fbd067bd36"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.379204 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.402713 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" event={"ID":"5819a852-b138-4647-bfeb-186464e53a17","Type":"ContainerStarted","Data":"9abd4551d908eb6729bc6a7b44b85f5af1c83d745beb2df204a520283319a7f0"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.408614 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" event={"ID":"748ee850-8101-42cd-b903-4c1443fa32e4","Type":"ContainerStarted","Data":"2dfb95407ab7fe980613666fb0a76b6ffdca943a145e63d446f116dc556a3ef6"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.420982 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" event={"ID":"18324d8e-bab7-4c65-9ad8-32878ffd7b0b","Type":"ContainerStarted","Data":"3f4cff62f902a2269db7587304d51109d4d8ec306040e1e1ea22b422bebc99a2"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.422961 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.423091 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.923070609 +0000 UTC m=+144.550160877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.423285 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.423757 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:02.923749171 +0000 UTC m=+144.550839439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.435180 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" event={"ID":"31ea7bfc-f519-411d-ad96-ab0edbbdac0f","Type":"ContainerStarted","Data":"25b2533fccac3959a9687a2ffb8406a3266d9905ea113ff37ca4bfe8901f9824"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.475683 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" event={"ID":"3f00b63a-3d97-443c-9613-e94b0cad6f9b","Type":"ContainerStarted","Data":"a1b692b8dd9203a6dd694deb34dcca6249552f805fdf819e4fea7aee9334f976"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.483022 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" event={"ID":"4d1c19bc-9031-49eb-88a1-3cb263134b3d","Type":"ContainerStarted","Data":"de50c513a7f22497734c80a4bd667b0ad2672c40f7385eed2474943958cc320f"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.484269 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.485962 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" event={"ID":"c67409be-2144-4b3d-be8c-81aaca215420","Type":"ContainerStarted","Data":"41d6e868599f1469012af809201ca11380d5b5f8a211b7f5853db7e4834b651a"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.507518 4675 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-plmm7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.507588 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" podUID="4d1c19bc-9031-49eb-88a1-3cb263134b3d" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.508158 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" event={"ID":"44bee603-4ff5-4319-aa19-d5737abba431","Type":"ContainerStarted","Data":"2379e325c836889b4427f63292ea05abda49a097b2eaba4ebd41d4a2ed9f16b3"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.523939 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.524098 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.024073801 +0000 UTC m=+144.651164079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.524368 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.528873 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.028853926 +0000 UTC m=+144.655944194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.546072 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6v99v"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.547871 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" event={"ID":"b6181566-7e6d-42a7-98cc-9cd735da6504","Type":"ContainerStarted","Data":"95389d6b0c6a4193c48fdf0d6efb6914921e2bbf2593c9d293483ebca6c441ef"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.547913 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" event={"ID":"b6181566-7e6d-42a7-98cc-9cd735da6504","Type":"ContainerStarted","Data":"aaae33dd83945105e4fc14e1176946700b9b545abfd484eef299eae126c1ea31"} Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.551192 4675 patch_prober.go:28] interesting pod/downloads-7954f5f757-pr5cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.551255 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pr5cf" podUID="cc0c9cff-8b28-4d9c-b50d-0ab81e16c841" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.569086 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.610887 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" podStartSLOduration=123.610851063 podStartE2EDuration="2m3.610851063s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:02.605690615 +0000 UTC m=+144.232780903" watchObservedRunningTime="2026-02-19 08:46:02.610851063 +0000 UTC m=+144.237941341" Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.625863 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.628079 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.12804814 +0000 UTC m=+144.755138468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.642891 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m2j8j"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.727417 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.727861 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.227846874 +0000 UTC m=+144.854937152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.785601 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-jzchz"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.838917 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.839694 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.339665687 +0000 UTC m=+144.966755965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.878171 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n"] Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.878300 4675 csr.go:261] certificate signing request csr-jcgvn is approved, waiting to be issued Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.912456 4675 csr.go:257] certificate signing request csr-jcgvn is issued Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.943355 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:02 crc kubenswrapper[4675]: E0219 08:46:02.952046 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.444332748 +0000 UTC m=+145.071423016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:02 crc kubenswrapper[4675]: I0219 08:46:02.990204 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-6nnjw" podStartSLOduration=123.990180314 podStartE2EDuration="2m3.990180314s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:02.987678303 +0000 UTC m=+144.614768591" watchObservedRunningTime="2026-02-19 08:46:02.990180314 +0000 UTC m=+144.617270582" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.050285 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.050934 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.550913062 +0000 UTC m=+145.178003330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.051377 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vg9l6" podStartSLOduration=124.051351306 podStartE2EDuration="2m4.051351306s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.02584499 +0000 UTC m=+144.652935258" watchObservedRunningTime="2026-02-19 08:46:03.051351306 +0000 UTC m=+144.678441574" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.079031 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-sdqq6"] Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.154287 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.156315 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-hhkkj" podStartSLOduration=124.156291956 podStartE2EDuration="2m4.156291956s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.144186614 +0000 UTC m=+144.771276882" watchObservedRunningTime="2026-02-19 08:46:03.156291956 +0000 UTC m=+144.783382224" Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.159848 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.659828181 +0000 UTC m=+145.286918449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.257802 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.258423 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.758400875 +0000 UTC m=+145.385491143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.336340 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fhvv2" podStartSLOduration=124.33631604 podStartE2EDuration="2m4.33631604s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.290359181 +0000 UTC m=+144.917449449" watchObservedRunningTime="2026-02-19 08:46:03.33631604 +0000 UTC m=+144.963406308" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.361440 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.361982 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.861964001 +0000 UTC m=+145.489054269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.398299 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk"] Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.407341 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7"] Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.463116 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.463494 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:03.96347433 +0000 UTC m=+145.590564598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.567200 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.568018 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.067995597 +0000 UTC m=+145.695085865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.582617 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" event={"ID":"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd","Type":"ContainerStarted","Data":"1880c29a43d8fae6c5cf8345bdb74b87bae1f13d36ff7b236fb1ac9ed332f789"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.596846 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" event={"ID":"3f00b63a-3d97-443c-9613-e94b0cad6f9b","Type":"ContainerStarted","Data":"3f8e401cb0e26dd1cf8f87b7c30334395a91ac9d9ce8a9f2e543603e75895ebb"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.611112 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6v99v" event={"ID":"914e8e63-cb4b-4784-84be-fa509224d2f1","Type":"ContainerStarted","Data":"79b51922e92e091103cde2fce969b39711fe56af3b393caea59f604de9cc1dbb"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.635514 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mrp4g" event={"ID":"24a4abb3-12f1-4f94-905b-344fc352730c","Type":"ContainerStarted","Data":"466cb6ea816674eba7737b7332d2b4ac90cbbaac5ca7cb38df27dece1e489cf8"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.635590 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mrp4g" event={"ID":"24a4abb3-12f1-4f94-905b-344fc352730c","Type":"ContainerStarted","Data":"f725813f0c13529234af9c6c66f2f1feef96a96c0d80bf4b49859bbc4848cf64"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.637718 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" event={"ID":"f0cbf82a-8336-4b7f-a96e-0329a0dffb40","Type":"ContainerStarted","Data":"17e3d9c020dc95ddc95e9f10af57c1a8718d93577740d3c158fefb2af0421969"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.640011 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" event={"ID":"5acabaa8-91d2-4f21-8a47-3959801a3600","Type":"ContainerStarted","Data":"7f1b0e39536f52e0dcccad7326ade76a1b99fcdf7c38057238acf0f5c419a442"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.658858 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" event={"ID":"cacb4d41-218a-441d-a62a-77338425aa3e","Type":"ContainerStarted","Data":"55d49b1fa41198d504e6c1523bb4980b9a501d75691ffc62c737b031ec461437"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.669266 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.670428 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" event={"ID":"ad9383f1-0293-4f42-b7f9-2e3fef0e4569","Type":"ContainerStarted","Data":"51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92"} Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.670673 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.170653213 +0000 UTC m=+145.797743481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.671325 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.687173 4675 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-srrlc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.687257 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" podUID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.689424 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ntvd2" podStartSLOduration=124.689403141 podStartE2EDuration="2m4.689403141s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.687784148 +0000 UTC m=+145.314874416" watchObservedRunningTime="2026-02-19 08:46:03.689403141 +0000 UTC m=+145.316493409" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.720345 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-rjqzp" event={"ID":"d3d57965-61a8-4a6f-883b-20360fd36509","Type":"ContainerStarted","Data":"bd55bef5f921d71f5644bcab065c446731338957ffaa6eb86f5712b039067316"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.733919 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" podStartSLOduration=124.733892602 podStartE2EDuration="2m4.733892602s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.733532211 +0000 UTC m=+145.360622479" watchObservedRunningTime="2026-02-19 08:46:03.733892602 +0000 UTC m=+145.360982860" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.750954 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" event={"ID":"c67409be-2144-4b3d-be8c-81aaca215420","Type":"ContainerStarted","Data":"9bd6770dbfcb85f217f2367c9aab4a892bbaba5dc5b826415226248db816923c"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.751008 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz"] Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.771914 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rpfbt" podStartSLOduration=124.771869033 podStartE2EDuration="2m4.771869033s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.764414341 +0000 UTC m=+145.391504619" watchObservedRunningTime="2026-02-19 08:46:03.771869033 +0000 UTC m=+145.398959301" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.773941 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.774281 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.274269171 +0000 UTC m=+145.901359429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.834869 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg"] Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.835299 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" event={"ID":"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f","Type":"ContainerStarted","Data":"02f8277cc51a19e8a99a586983263a96cfc2ad2cf30573c36ca873d774a4e690"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.851533 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tf9mf" podStartSLOduration=124.851506823 podStartE2EDuration="2m4.851506823s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.803696824 +0000 UTC m=+145.430787102" watchObservedRunningTime="2026-02-19 08:46:03.851506823 +0000 UTC m=+145.478597091" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.851693 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cwrvv"] Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.864183 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" podStartSLOduration=124.862365935 podStartE2EDuration="2m4.862365935s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.855903066 +0000 UTC m=+145.482993334" watchObservedRunningTime="2026-02-19 08:46:03.862365935 +0000 UTC m=+145.489456203" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.876708 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.877073 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.377055531 +0000 UTC m=+146.004145799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.877124 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" event={"ID":"a7fb92b9-e97b-46a0-b3da-c9186bd4b585","Type":"ContainerStarted","Data":"74237a149a1a834ca9622291382627efa4efd5165f8fc7d0ad210f3108714096"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.882448 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" event={"ID":"230232aa-5a5d-4436-a495-ecd2919fa56b","Type":"ContainerStarted","Data":"0467c31780369465cf3bb90a1706ee9b07d78ebf746d40340e81ff323a8c82b1"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.885742 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" podStartSLOduration=124.885708881 podStartE2EDuration="2m4.885708881s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:03.882740345 +0000 UTC m=+145.509830613" watchObservedRunningTime="2026-02-19 08:46:03.885708881 +0000 UTC m=+145.512799149" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.893266 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" event={"ID":"44bee603-4ff5-4319-aa19-d5737abba431","Type":"ContainerStarted","Data":"3a48930ec95d706f0863fecce904c4cc1008e1ee529b25631d8de48064b6bd78"} Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.909570 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.914689 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-19 08:41:02 +0000 UTC, rotation deadline is 2027-01-12 07:28:40.516101004 +0000 UTC Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.914725 4675 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7846h42m36.601378312s for next certificate rotation Feb 19 08:46:03 crc kubenswrapper[4675]: W0219 08:46:03.920925 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod816b48eb_9013_414f_a400_909963ff61fe.slice/crio-a5c7fc718a564cbbb1874881ab08fbe41b12d119d865691fdb107d41df038f06 WatchSource:0}: Error finding container a5c7fc718a564cbbb1874881ab08fbe41b12d119d865691fdb107d41df038f06: Status 404 returned error can't find the container with id a5c7fc718a564cbbb1874881ab08fbe41b12d119d865691fdb107d41df038f06 Feb 19 08:46:03 crc kubenswrapper[4675]: I0219 08:46:03.977996 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:03 crc kubenswrapper[4675]: E0219 08:46:03.985752 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.485732012 +0000 UTC m=+146.112822330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.018139 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp"] Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.076974 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5b9t"] Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.090376 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.090850 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.590827608 +0000 UTC m=+146.217917876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.095619 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5l26w"] Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.098995 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dj8qx" podStartSLOduration=125.098974492 podStartE2EDuration="2m5.098974492s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:04.081206756 +0000 UTC m=+145.708297024" watchObservedRunningTime="2026-02-19 08:46:04.098974492 +0000 UTC m=+145.726064760" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.118712 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" podStartSLOduration=125.11868058 podStartE2EDuration="2m5.11868058s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:04.113949507 +0000 UTC m=+145.741039795" watchObservedRunningTime="2026-02-19 08:46:04.11868058 +0000 UTC m=+145.745770848" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.138910 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45"] Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.181959 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67"] Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.192652 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.193305 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.693291628 +0000 UTC m=+146.320381896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.259866 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-rjqzp" podStartSLOduration=125.259842604 podStartE2EDuration="2m5.259842604s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:04.259265206 +0000 UTC m=+145.886355474" watchObservedRunningTime="2026-02-19 08:46:04.259842604 +0000 UTC m=+145.886932872" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.294065 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.294233 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.794200608 +0000 UTC m=+146.421290886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.294446 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.294852 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.794834718 +0000 UTC m=+146.421924986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.382012 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-mrp4g" podStartSLOduration=6.381991252 podStartE2EDuration="6.381991252s" podCreationTimestamp="2026-02-19 08:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:04.380352209 +0000 UTC m=+146.007442497" watchObservedRunningTime="2026-02-19 08:46:04.381991252 +0000 UTC m=+146.009081520" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.396483 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.396888 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:04.896872735 +0000 UTC m=+146.523963003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.432357 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" podStartSLOduration=125.432317743 podStartE2EDuration="2m5.432317743s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:04.42758453 +0000 UTC m=+146.054674798" watchObservedRunningTime="2026-02-19 08:46:04.432317743 +0000 UTC m=+146.059408011" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.497944 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.505918 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.005890717 +0000 UTC m=+146.632980985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.511714 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rc9fc" podStartSLOduration=125.511679635 podStartE2EDuration="2m5.511679635s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:04.505790144 +0000 UTC m=+146.132880412" watchObservedRunningTime="2026-02-19 08:46:04.511679635 +0000 UTC m=+146.138769903" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.599821 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.600251 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.100232924 +0000 UTC m=+146.727323192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.646123 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.652596 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:04 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:04 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:04 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.652678 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.701085 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.701472 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.201458234 +0000 UTC m=+146.828548502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.803415 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.803847 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.303828961 +0000 UTC m=+146.930919229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.907298 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:04 crc kubenswrapper[4675]: E0219 08:46:04.907965 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.407952995 +0000 UTC m=+147.035043263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.990361 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" event={"ID":"816b48eb-9013-414f-a400-909963ff61fe","Type":"ContainerStarted","Data":"5455ced94002748a140f9e952a39f8155a9635e22244ee9c162a816337d5bcd0"} Feb 19 08:46:04 crc kubenswrapper[4675]: I0219 08:46:04.990458 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" event={"ID":"816b48eb-9013-414f-a400-909963ff61fe","Type":"ContainerStarted","Data":"a5c7fc718a564cbbb1874881ab08fbe41b12d119d865691fdb107d41df038f06"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.010270 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.011745 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.511723627 +0000 UTC m=+147.138813895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.014650 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rj8cg" podStartSLOduration=126.014620451 podStartE2EDuration="2m6.014620451s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.013531866 +0000 UTC m=+146.640622154" watchObservedRunningTime="2026-02-19 08:46:05.014620451 +0000 UTC m=+146.641710719" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.032847 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" event={"ID":"f0cbf82a-8336-4b7f-a96e-0329a0dffb40","Type":"ContainerStarted","Data":"96f790519d6b8170a140f9fa3d269c67c18304cc735ae5359dbf6387d5064a16"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.032975 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" event={"ID":"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd","Type":"ContainerStarted","Data":"34ac4eed0ec6be49c9adb1fff2a6b576a153d33e326eb80ac1eeb348009129c1"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.034515 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6v99v" event={"ID":"914e8e63-cb4b-4784-84be-fa509224d2f1","Type":"ContainerStarted","Data":"2d0b7fa10c089c0a356d0a23724e6e759b3a9d7d98e728c71ab0a1c1584dfa12"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.040931 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" event={"ID":"d4011782-4d60-4178-9ed5-e76a79a672aa","Type":"ContainerStarted","Data":"88fe20785efb1e063198dab48e4a97a9f8de737367c3514516712f8d40b5d8e3"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.040998 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" event={"ID":"d4011782-4d60-4178-9ed5-e76a79a672aa","Type":"ContainerStarted","Data":"845459d3aa0760231534afe3da508d32ab0986cdced93e30e9b6faed67e9b60c"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.043236 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" event={"ID":"31ea7bfc-f519-411d-ad96-ab0edbbdac0f","Type":"ContainerStarted","Data":"fcdf8bd606d17610bb3a59d1169a75d5d929f7f33744b83423c876d922a87304"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.045379 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" event={"ID":"748ee850-8101-42cd-b903-4c1443fa32e4","Type":"ContainerStarted","Data":"b51457dcc59257248b0f9e5b059eaf09789342927817df425946898dd5bfe114"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.066458 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" event={"ID":"f2da5f24-2b80-4967-8a86-ab370a38e45f","Type":"ContainerStarted","Data":"3c6efcfd7c04f60223cfc2614b775da372aae36ab83016bdf709677383fc6723"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.074074 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" event={"ID":"86e4bad1-6678-41f1-bd97-f48fc5312b10","Type":"ContainerStarted","Data":"6b50b22ef88a75adf07d48f386a225a7eb43b1595d4ecbb7850adde2b856545d"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.074129 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" event={"ID":"86e4bad1-6678-41f1-bd97-f48fc5312b10","Type":"ContainerStarted","Data":"9b13f9b74f93f902480dbdf679d81d35b88a82554d8e5ef3831f41ea5a316599"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.075174 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.085455 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" event={"ID":"5acabaa8-91d2-4f21-8a47-3959801a3600","Type":"ContainerStarted","Data":"c2e90d25828ff9b7c1e70720f4b442d24c4ada653404156737357919a9444725"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.094999 4675 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-p8bgp container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.095071 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" podUID="86e4bad1-6678-41f1-bd97-f48fc5312b10" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.102172 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" event={"ID":"bbed8b9a-ef34-424b-81e0-457a38384ba0","Type":"ContainerStarted","Data":"28490fc99df0a57f9d4d004f3715d51b5bbf05b680ff4769226fac46d99491e0"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.104895 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-86dwp" podStartSLOduration=126.104859725 podStartE2EDuration="2m6.104859725s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.097320341 +0000 UTC m=+146.724410619" watchObservedRunningTime="2026-02-19 08:46:05.104859725 +0000 UTC m=+146.731949993" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.141775 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vmq9z" podStartSLOduration=126.141750051 podStartE2EDuration="2m6.141750051s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.129554356 +0000 UTC m=+146.756644644" watchObservedRunningTime="2026-02-19 08:46:05.141750051 +0000 UTC m=+146.768840319" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.146173 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" event={"ID":"bbed8b9a-ef34-424b-81e0-457a38384ba0","Type":"ContainerStarted","Data":"578c3601be1878dc00b4e1b72e64139f3381f3c7ffa2e714f2080d831f975634"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.149365 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" event={"ID":"d35a2814-4eb0-4e49-be8c-bbde18573d00","Type":"ContainerStarted","Data":"a0e3a7af67cd858a9703a4390ad70b56fb792c6954b6d0fce6dfca631a822b2b"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.149856 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.158949 4675 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wtr45 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.159042 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" podUID="d35a2814-4eb0-4e49-be8c-bbde18573d00" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.163431 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.173269 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.673165409 +0000 UTC m=+147.300255677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.182986 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-sdqq6" podStartSLOduration=126.182958496 podStartE2EDuration="2m6.182958496s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.174324786 +0000 UTC m=+146.801415054" watchObservedRunningTime="2026-02-19 08:46:05.182958496 +0000 UTC m=+146.810048764" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.186531 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" event={"ID":"230232aa-5a5d-4436-a495-ecd2919fa56b","Type":"ContainerStarted","Data":"3195dbb00b995c41d12ab32fafd8e739cc97bf55ed15792db7f9241b00694ac8"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.210342 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.221458 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.225556 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" event={"ID":"24727689-f73c-4ff3-9b0c-9fa625d16c27","Type":"ContainerStarted","Data":"43823c8f0b73b1880076f8b6f2a3bf6d28784f0ac352c996d769218af290fee3"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.225948 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" event={"ID":"24727689-f73c-4ff3-9b0c-9fa625d16c27","Type":"ContainerStarted","Data":"06bab7dc58d1e87edb1e04686a9a2c08f489276d7bed502ec9b82cf2036a7553"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.226721 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.238989 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.252606 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b8w5h" podStartSLOduration=126.252582302 podStartE2EDuration="2m6.252582302s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.210160747 +0000 UTC m=+146.837251005" watchObservedRunningTime="2026-02-19 08:46:05.252582302 +0000 UTC m=+146.879672570" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.264487 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" event={"ID":"cacb4d41-218a-441d-a62a-77338425aa3e","Type":"ContainerStarted","Data":"d058c8b3733e8a9e6297aaa85715ce069400fb082f5943eec5709eb52bbaaad2"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.268587 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" event={"ID":"75b0c6fb-238c-4f52-927e-920b288a5f85","Type":"ContainerStarted","Data":"e515ba7a8aa8d7f53952bf3b47c4ce342b1e8842e5c9dc41903e9c2b2e45b8b8"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.268663 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" event={"ID":"75b0c6fb-238c-4f52-927e-920b288a5f85","Type":"ContainerStarted","Data":"77aef01343cf0d8efd11f5eb2644755ed8831bc7ffe409b095785e2d5e2024d1"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.270438 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" event={"ID":"f8c1b4c7-ae04-4eb0-85e7-a8daa586cd1f","Type":"ContainerStarted","Data":"635d143e0d6fd4801d0ffb7ea1ee9b8956018314cc88b45469338112799bb2d7"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.272629 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" event={"ID":"02a5e839-abbb-424e-b702-7e197d9d3307","Type":"ContainerStarted","Data":"cc189ccec64a3bc645c57411de03ecaee427d376506bd560c0ca400d494881d0"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.273846 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.274839 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.774821903 +0000 UTC m=+147.401912171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.276846 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" event={"ID":"5819a852-b138-4647-bfeb-186464e53a17","Type":"ContainerStarted","Data":"84d433809fece5a40384431bcfada4cdbda046db75263753fb76d50fe241de20"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.276887 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" event={"ID":"5819a852-b138-4647-bfeb-186464e53a17","Type":"ContainerStarted","Data":"09b58a9ebd9b106c92d36e8896abfb916a3c600ecf32c63c61233f761602a7b7"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.278913 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" podStartSLOduration=126.278897855 podStartE2EDuration="2m6.278897855s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.25314325 +0000 UTC m=+146.880233518" watchObservedRunningTime="2026-02-19 08:46:05.278897855 +0000 UTC m=+146.905988123" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.285024 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.285598 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.305034 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5l26w" event={"ID":"a80cf4f9-94b9-4694-85e5-9643aace2211","Type":"ContainerStarted","Data":"67fc5f652f8b36a058cef3f66b1f30f0a789fc0eb15f0ee682e18f485cd47ea5"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.305106 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5l26w" event={"ID":"a80cf4f9-94b9-4694-85e5-9643aace2211","Type":"ContainerStarted","Data":"f4ed8989de776878dd0a664ff8a24b84e433aa7f703f65e8f0d5f26c46276bc9"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.316124 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xrzvw" podStartSLOduration=126.31610408 podStartE2EDuration="2m6.31610408s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.314765667 +0000 UTC m=+146.941855945" watchObservedRunningTime="2026-02-19 08:46:05.31610408 +0000 UTC m=+146.943194348" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.316762 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" podStartSLOduration=65.316756561 podStartE2EDuration="1m5.316756561s" podCreationTimestamp="2026-02-19 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.279951889 +0000 UTC m=+146.907042157" watchObservedRunningTime="2026-02-19 08:46:05.316756561 +0000 UTC m=+146.943846829" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.324513 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-jzchz" event={"ID":"a7fb92b9-e97b-46a0-b3da-c9186bd4b585","Type":"ContainerStarted","Data":"311ecb135fabe859ec11d0b329fa78555467e02c7680700ba76977146cc6299e"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.338198 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" event={"ID":"017fb59c-7331-42bf-9a21-8dff74814740","Type":"ContainerStarted","Data":"890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.338238 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" event={"ID":"017fb59c-7331-42bf-9a21-8dff74814740","Type":"ContainerStarted","Data":"59a9ee2a13ded4c16f09461c51fc18c8db80bbe4a494f8aeddc395e53cdebfda"} Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.338253 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.340173 4675 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cwrvv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.340320 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.341002 4675 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-srrlc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.341099 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" podUID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.345627 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" podStartSLOduration=126.345570815 podStartE2EDuration="2m6.345570815s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.343373854 +0000 UTC m=+146.970464122" watchObservedRunningTime="2026-02-19 08:46:05.345570815 +0000 UTC m=+146.972661103" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.353172 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-99sr2" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.356786 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rt65v" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.378473 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.378925 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" podStartSLOduration=126.378912365 podStartE2EDuration="2m6.378912365s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.37750076 +0000 UTC m=+147.004591028" watchObservedRunningTime="2026-02-19 08:46:05.378912365 +0000 UTC m=+147.006002633" Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.383064 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.883046149 +0000 UTC m=+147.510136407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.442118 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s8cwp" podStartSLOduration=126.442090073 podStartE2EDuration="2m6.442090073s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.421538047 +0000 UTC m=+147.048628325" watchObservedRunningTime="2026-02-19 08:46:05.442090073 +0000 UTC m=+147.069180331" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.462263 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b9t9z" podStartSLOduration=126.462240085 podStartE2EDuration="2m6.462240085s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.461340956 +0000 UTC m=+147.088431224" watchObservedRunningTime="2026-02-19 08:46:05.462240085 +0000 UTC m=+147.089330353" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.482595 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.483087 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:05.98306468 +0000 UTC m=+147.610154938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.544443 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wf8zz" podStartSLOduration=126.544410118 podStartE2EDuration="2m6.544410118s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.512024299 +0000 UTC m=+147.139114557" watchObservedRunningTime="2026-02-19 08:46:05.544410118 +0000 UTC m=+147.171500386" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.586469 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.587090 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.08705869 +0000 UTC m=+147.714148958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.637522 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" podStartSLOduration=126.637492784 podStartE2EDuration="2m6.637492784s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.604262337 +0000 UTC m=+147.231352605" watchObservedRunningTime="2026-02-19 08:46:05.637492784 +0000 UTC m=+147.264583052" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.653910 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:05 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:05 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:05 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.653996 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.690197 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.690810 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.190786531 +0000 UTC m=+147.817876799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.711909 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" podStartSLOduration=126.711884075 podStartE2EDuration="2m6.711884075s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.710548771 +0000 UTC m=+147.337639039" watchObservedRunningTime="2026-02-19 08:46:05.711884075 +0000 UTC m=+147.338974353" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.791779 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.792264 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.292231528 +0000 UTC m=+147.919321796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.893351 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.893604 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.393565862 +0000 UTC m=+148.020656130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.893873 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.894334 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.394314586 +0000 UTC m=+148.021404884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.968137 4675 patch_prober.go:28] interesting pod/apiserver-76f77b778f-sdrxk container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]log ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]etcd ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/generic-apiserver-start-informers ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/max-in-flight-filter ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 19 08:46:05 crc kubenswrapper[4675]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 19 08:46:05 crc kubenswrapper[4675]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/project.openshift.io-projectcache ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/openshift.io-startinformers ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 19 08:46:05 crc kubenswrapper[4675]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 19 08:46:05 crc kubenswrapper[4675]: livez check failed Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.968249 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" podUID="cacb4d41-218a-441d-a62a-77338425aa3e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:05 crc kubenswrapper[4675]: I0219 08:46:05.995253 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:05 crc kubenswrapper[4675]: E0219 08:46:05.996169 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.496147516 +0000 UTC m=+148.123237784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.097710 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.098202 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.598183701 +0000 UTC m=+148.225273979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.198788 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.199021 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.698984097 +0000 UTC m=+148.326074375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.199126 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.199576 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.699565646 +0000 UTC m=+148.326655994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.300291 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.800269339 +0000 UTC m=+148.427359607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.300166 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.300545 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.301829 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.301902 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.301951 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.301983 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.302346 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.802333366 +0000 UTC m=+148.429423634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.302854 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.315005 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.328859 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.330405 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.342558 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.343791 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" event={"ID":"d35a2814-4eb0-4e49-be8c-bbde18573d00","Type":"ContainerStarted","Data":"7784ca5dc9b2c89df1dabda1f92893d8abbd53738f0ae5328d2d6d9dfa099007"} Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.344764 4675 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wtr45 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.344901 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" podUID="d35a2814-4eb0-4e49-be8c-bbde18573d00" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.348227 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" event={"ID":"230232aa-5a5d-4436-a495-ecd2919fa56b","Type":"ContainerStarted","Data":"d864fd8fc0e951ff2dd9357c4f33a4defd546dc6791e4c422fb9f6b7903919ae"} Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.350778 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" event={"ID":"d4011782-4d60-4178-9ed5-e76a79a672aa","Type":"ContainerStarted","Data":"b69fd1211f2a10a0759a51eeab644329eb4ae5866a05dd690bb218882ffb0a22"} Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.354280 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" event={"ID":"24727689-f73c-4ff3-9b0c-9fa625d16c27","Type":"ContainerStarted","Data":"899a713484ebe31dacc78a8208fc6c4bd34bfdee893da2f341f335b5ebf50266"} Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.356289 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" event={"ID":"f2da5f24-2b80-4967-8a86-ab370a38e45f","Type":"ContainerStarted","Data":"1926a25b32e5bb7228ec29a399e9ef5e2c8d4f3898620bd46a388dcc14438fc1"} Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.356359 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.360705 4675 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d9r67 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.361126 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" podUID="f2da5f24-2b80-4967-8a86-ab370a38e45f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.371367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" event={"ID":"02a5e839-abbb-424e-b702-7e197d9d3307","Type":"ContainerStarted","Data":"402fc5a19c17e8b0380d838f0004d39f7ba8d29b6502e97f9d2967408a7eeed8"} Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.391610 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-5l26w" podStartSLOduration=8.391586088 podStartE2EDuration="8.391586088s" podCreationTimestamp="2026-02-19 08:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:05.747888011 +0000 UTC m=+147.374978289" watchObservedRunningTime="2026-02-19 08:46:06.391586088 +0000 UTC m=+148.018676356" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.393038 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-m2j8j" podStartSLOduration=127.393027795 podStartE2EDuration="2m7.393027795s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:06.391804095 +0000 UTC m=+148.018894363" watchObservedRunningTime="2026-02-19 08:46:06.393027795 +0000 UTC m=+148.020118063" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.402255 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6v99v" event={"ID":"914e8e63-cb4b-4784-84be-fa509224d2f1","Type":"ContainerStarted","Data":"40d251e35e6cd4c492e2a24c35e3a209a7be147176c38cae8e38cd793eaa4d85"} Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.403108 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.403681 4675 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cwrvv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.403730 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.403829 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:06.903806614 +0000 UTC m=+148.530896932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.426752 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.433105 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-p8bgp" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.443910 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" podStartSLOduration=127.443891333 podStartE2EDuration="2m7.443891333s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:06.443404277 +0000 UTC m=+148.070494565" watchObservedRunningTime="2026-02-19 08:46:06.443891333 +0000 UTC m=+148.070981601" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.505444 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.507092 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.00707333 +0000 UTC m=+148.634163588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.535714 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t6vv7" podStartSLOduration=127.535694957 podStartE2EDuration="2m7.535694957s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:06.475070023 +0000 UTC m=+148.102160311" watchObservedRunningTime="2026-02-19 08:46:06.535694957 +0000 UTC m=+148.162785225" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.551269 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.578534 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6v99v" podStartSLOduration=8.578515235 podStartE2EDuration="8.578515235s" podCreationTimestamp="2026-02-19 08:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:06.535841892 +0000 UTC m=+148.162932160" watchObservedRunningTime="2026-02-19 08:46:06.578515235 +0000 UTC m=+148.205605493" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.629780 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.630368 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.632092 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.13207518 +0000 UTC m=+148.759165448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.655048 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:06 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:06 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:06 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.655167 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.733074 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.733907 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.2338922 +0000 UTC m=+148.860982468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.834414 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.834872 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.334854261 +0000 UTC m=+148.961944529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:06 crc kubenswrapper[4675]: I0219 08:46:06.936611 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:06 crc kubenswrapper[4675]: E0219 08:46:06.937018 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.436993031 +0000 UTC m=+149.064083299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.037402 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.038547 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.53849979 +0000 UTC m=+149.165590098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.145374 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.145935 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.64591836 +0000 UTC m=+149.273008638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.246417 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.246761 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.746743107 +0000 UTC m=+149.373833375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.347580 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.348496 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.848477404 +0000 UTC m=+149.475567672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.398914 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.399611 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.412136 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.412470 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.416597 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.433431 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5bdfa3fbc88d492bda29ce49c9c4c45748937417f2bb97437e48cd9451a68c9a"} Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.450628 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.451029 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:07.951008886 +0000 UTC m=+149.578099154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.489045 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e2153de8d5d31ad1a94cea27306eb95056187cd09c93951e23e568a2d05e98f8"} Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.491463 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.520422 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wtr45" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.553856 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.554218 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.554353 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.555140 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.05512787 +0000 UTC m=+149.682218138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.647578 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:07 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:07 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:07 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.647759 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.656331 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.656734 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.656836 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.156794504 +0000 UTC m=+149.783884932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.657100 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.659398 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.159380228 +0000 UTC m=+149.786470686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.663809 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.671931 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.706981 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.733030 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.765952 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.766274 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.266256931 +0000 UTC m=+149.893347199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.867196 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.867571 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.367557504 +0000 UTC m=+149.994647772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.968476 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.968722 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.46868761 +0000 UTC m=+150.095777878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:07 crc kubenswrapper[4675]: I0219 08:46:07.968881 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:07 crc kubenswrapper[4675]: E0219 08:46:07.969255 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.469238358 +0000 UTC m=+150.096328626 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.070624 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.071020 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.571003056 +0000 UTC m=+150.198093324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.174574 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.175123 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.675102149 +0000 UTC m=+150.302192417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.277388 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.277616 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.777578059 +0000 UTC m=+150.404668327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.277863 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.278284 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.778278072 +0000 UTC m=+150.405368330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.379218 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9r67" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.379284 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.379393 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.879370858 +0000 UTC m=+150.506461126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.379768 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.380208 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.880177904 +0000 UTC m=+150.507268172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.395932 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 19 08:46:08 crc kubenswrapper[4675]: W0219 08:46:08.401719 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1bbe28ee_f09f_45c1_aae3_fa37811776f2.slice/crio-99428a340e7d29a16eaf648b3e60eddb411611bad1a0c9bcd7502d61412869a6 WatchSource:0}: Error finding container 99428a340e7d29a16eaf648b3e60eddb411611bad1a0c9bcd7502d61412869a6: Status 404 returned error can't find the container with id 99428a340e7d29a16eaf648b3e60eddb411611bad1a0c9bcd7502d61412869a6 Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.432946 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vc6hh"] Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.434016 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.436530 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.460338 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vc6hh"] Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.480336 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.480758 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:08.980737232 +0000 UTC m=+150.607827500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.512221 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" event={"ID":"02a5e839-abbb-424e-b702-7e197d9d3307","Type":"ContainerStarted","Data":"b053ad529d3f98343a3900d6fc9c84464c82cf5f0ff0e2ce2344803e13e29afa"} Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.512276 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" event={"ID":"02a5e839-abbb-424e-b702-7e197d9d3307","Type":"ContainerStarted","Data":"b3981ea0b1580324840c420fdf0ef94f61433ed7fcd0104fd61421aa86bdcb8c"} Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.519457 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1bbe28ee-f09f-45c1-aae3-fa37811776f2","Type":"ContainerStarted","Data":"99428a340e7d29a16eaf648b3e60eddb411611bad1a0c9bcd7502d61412869a6"} Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.521625 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2a92f76d15224bcad4090aa34f40f523a358e5a8d117691e0da6d70759f7cd49"} Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.532493 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7c9fdcd43871dc9db072764e9a1dc3f30a5145cbc5451c0fa169453697d3d210"} Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.532543 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0e6d8f35c0432640a5c727be4e3aa8a1cb764d167b651fe878b14f5f04916a80"} Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.556553 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"aaa9f2357f7636276111eeaca0b4e0b4552abec7d8b244451d71da2c2faf81dd"} Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.557688 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.585565 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pskkw\" (UniqueName: \"kubernetes.io/projected/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-kube-api-access-pskkw\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.585686 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-catalog-content\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.585846 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.585896 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-utilities\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.586279 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.086260012 +0000 UTC m=+150.713350320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.625961 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vbjkd"] Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.628564 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.635897 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.647230 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbjkd"] Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.651298 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:08 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:08 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:08 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.651371 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.690289 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.690626 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pskkw\" (UniqueName: \"kubernetes.io/projected/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-kube-api-access-pskkw\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.690718 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-catalog-content\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.690849 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-utilities\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.691544 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-catalog-content\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.691728 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.191701518 +0000 UTC m=+150.818791866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.693435 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-utilities\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.748870 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pskkw\" (UniqueName: \"kubernetes.io/projected/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-kube-api-access-pskkw\") pod \"community-operators-vc6hh\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.776201 4675 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.779015 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.792389 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-utilities\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.792451 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-catalog-content\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.792497 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz852\" (UniqueName: \"kubernetes.io/projected/52c2e6ec-6246-4346-8d44-f9d37b805bb0-kube-api-access-xz852\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.792532 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.793039 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.293023571 +0000 UTC m=+150.920113889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.806175 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t7b6d"] Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.810380 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.814473 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7b6d"] Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.896441 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.896885 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-utilities\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.896918 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-catalog-content\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.896943 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz852\" (UniqueName: \"kubernetes.io/projected/52c2e6ec-6246-4346-8d44-f9d37b805bb0-kube-api-access-xz852\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: E0219 08:46:08.897336 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.397320201 +0000 UTC m=+151.024410469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.897944 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-utilities\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.898156 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-catalog-content\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.927492 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz852\" (UniqueName: \"kubernetes.io/projected/52c2e6ec-6246-4346-8d44-f9d37b805bb0-kube-api-access-xz852\") pod \"certified-operators-vbjkd\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:08 crc kubenswrapper[4675]: I0219 08:46:08.985292 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.000993 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-utilities\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.001068 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzjmh\" (UniqueName: \"kubernetes.io/projected/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-kube-api-access-xzjmh\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.001101 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-catalog-content\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.001149 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:09 crc kubenswrapper[4675]: E0219 08:46:09.001576 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.501555608 +0000 UTC m=+151.128645876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.005281 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kk7x5"] Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.006672 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.039971 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kk7x5"] Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.102061 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.102312 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-utilities\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.102365 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzjmh\" (UniqueName: \"kubernetes.io/projected/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-kube-api-access-xzjmh\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.102393 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-catalog-content\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.102889 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-catalog-content\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.103034 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-utilities\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: E0219 08:46:09.103234 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.603190641 +0000 UTC m=+151.230280999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.158814 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzjmh\" (UniqueName: \"kubernetes.io/projected/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-kube-api-access-xzjmh\") pod \"community-operators-t7b6d\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.185557 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vc6hh"] Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.204281 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.205612 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-catalog-content\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.205677 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-utilities\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.205758 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.205790 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm4j6\" (UniqueName: \"kubernetes.io/projected/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-kube-api-access-gm4j6\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: E0219 08:46:09.206108 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.706096016 +0000 UTC m=+151.333186284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jfck7" (UID: "08306565-7dde-46be-b2de-d813dfdd976a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.281758 4675 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-19T08:46:08.776239587Z","Handler":null,"Name":""} Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.311349 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.312108 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm4j6\" (UniqueName: \"kubernetes.io/projected/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-kube-api-access-gm4j6\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.312161 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-catalog-content\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.312190 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-utilities\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.312705 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-utilities\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: E0219 08:46:09.312788 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-19 08:46:09.812767172 +0000 UTC m=+151.439857440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.313351 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-catalog-content\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.350558 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm4j6\" (UniqueName: \"kubernetes.io/projected/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-kube-api-access-gm4j6\") pod \"certified-operators-kk7x5\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.386002 4675 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.386063 4675 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.413307 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.419466 4675 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.419557 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.457662 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jfck7\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.460222 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbjkd"] Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.514331 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.520276 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.561687 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc6hh" event={"ID":"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2","Type":"ContainerStarted","Data":"69dca1d13c27bde045bd6c04f3c3862002065ed38d9514273ff7de814388c5a9"} Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.570924 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" event={"ID":"02a5e839-abbb-424e-b702-7e197d9d3307","Type":"ContainerStarted","Data":"e759fbb086523ad7f3ac106a42577163675aafbd45897058e9d406dc28528656"} Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.575559 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbjkd" event={"ID":"52c2e6ec-6246-4346-8d44-f9d37b805bb0","Type":"ContainerStarted","Data":"e4917ffa058b7f8e2b701a48eb05eb9617ddcc71671745cf03b02c219403f5a1"} Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.593676 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-n5b9t" podStartSLOduration=11.593655284 podStartE2EDuration="11.593655284s" podCreationTimestamp="2026-02-19 08:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:09.591136212 +0000 UTC m=+151.218226480" watchObservedRunningTime="2026-02-19 08:46:09.593655284 +0000 UTC m=+151.220745562" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.600113 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1bbe28ee-f09f-45c1-aae3-fa37811776f2","Type":"ContainerStarted","Data":"6c087c6362dbae6d07d6f265a00659556f6a673ca5628661b42cc4827d724301"} Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.637753 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.638461 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7b6d"] Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.645246 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:09 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:09 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:09 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.645548 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.685736 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.847299 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kk7x5"] Feb 19 08:46:09 crc kubenswrapper[4675]: W0219 08:46:09.859409 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd1115a1_2835_4bc5_8d17_7387b7f34cc4.slice/crio-7d6fef11f2d7c48beb46e0e844fb2a8d247db06d853fe11152f110ece52f7392 WatchSource:0}: Error finding container 7d6fef11f2d7c48beb46e0e844fb2a8d247db06d853fe11152f110ece52f7392: Status 404 returned error can't find the container with id 7d6fef11f2d7c48beb46e0e844fb2a8d247db06d853fe11152f110ece52f7392 Feb 19 08:46:09 crc kubenswrapper[4675]: I0219 08:46:09.922204 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jfck7"] Feb 19 08:46:09 crc kubenswrapper[4675]: W0219 08:46:09.950667 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08306565_7dde_46be_b2de_d813dfdd976a.slice/crio-98cd179ce5898a8d4c3bb46a00f1110934a579e96fccd4c949f1c04282a3a007 WatchSource:0}: Error finding container 98cd179ce5898a8d4c3bb46a00f1110934a579e96fccd4c949f1c04282a3a007: Status 404 returned error can't find the container with id 98cd179ce5898a8d4c3bb46a00f1110934a579e96fccd4c949f1c04282a3a007 Feb 19 08:46:09 crc kubenswrapper[4675]: E0219 08:46:09.974287 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52c2e6ec_6246_4346_8d44_f9d37b805bb0.slice/crio-conmon-4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e8bfbea_83ba_43fd_a0e6_1af01b28cbc4.slice/crio-conmon-440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7.scope\": RecentStats: unable to find data in memory cache]" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.127692 4675 patch_prober.go:28] interesting pod/downloads-7954f5f757-pr5cf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.127780 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-pr5cf" podUID="cc0c9cff-8b28-4d9c-b50d-0ab81e16c841" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.127719 4675 patch_prober.go:28] interesting pod/downloads-7954f5f757-pr5cf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.127873 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pr5cf" podUID="cc0c9cff-8b28-4d9c-b50d-0ab81e16c841" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.272503 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.272585 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.273979 4675 patch_prober.go:28] interesting pod/console-f9d7485db-vg9l6 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.274159 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vg9l6" podUID="0b433b98-d626-4286-a89b-3618ba63b58a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.288944 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.295149 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-sdrxk" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.602802 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sgxzf"] Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.604060 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.633227 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.636886 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" event={"ID":"08306565-7dde-46be-b2de-d813dfdd976a","Type":"ContainerStarted","Data":"20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.636942 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" event={"ID":"08306565-7dde-46be-b2de-d813dfdd976a","Type":"ContainerStarted","Data":"98cd179ce5898a8d4c3bb46a00f1110934a579e96fccd4c949f1c04282a3a007"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.637998 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.642834 4675 generic.go:334] "Generic (PLEG): container finished" podID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerID="d25eaae249951e387b7e626a1280c9c77de5bc0131d6bf83aa18a6ccce8aaad1" exitCode=0 Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.643162 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kk7x5" event={"ID":"fd1115a1-2835-4bc5-8d17-7387b7f34cc4","Type":"ContainerDied","Data":"d25eaae249951e387b7e626a1280c9c77de5bc0131d6bf83aa18a6ccce8aaad1"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.643316 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kk7x5" event={"ID":"fd1115a1-2835-4bc5-8d17-7387b7f34cc4","Type":"ContainerStarted","Data":"7d6fef11f2d7c48beb46e0e844fb2a8d247db06d853fe11152f110ece52f7392"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.653148 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:10 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:10 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:10 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.653229 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.653422 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.657143 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgxzf"] Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.660633 4675 generic.go:334] "Generic (PLEG): container finished" podID="e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" containerID="34ac4eed0ec6be49c9adb1fff2a6b576a153d33e326eb80ac1eeb348009129c1" exitCode=0 Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.660762 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" event={"ID":"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd","Type":"ContainerDied","Data":"34ac4eed0ec6be49c9adb1fff2a6b576a153d33e326eb80ac1eeb348009129c1"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.664477 4675 generic.go:334] "Generic (PLEG): container finished" podID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerID="4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d" exitCode=0 Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.664569 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbjkd" event={"ID":"52c2e6ec-6246-4346-8d44-f9d37b805bb0","Type":"ContainerDied","Data":"4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.708893 4675 generic.go:334] "Generic (PLEG): container finished" podID="1bbe28ee-f09f-45c1-aae3-fa37811776f2" containerID="6c087c6362dbae6d07d6f265a00659556f6a673ca5628661b42cc4827d724301" exitCode=0 Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.709289 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1bbe28ee-f09f-45c1-aae3-fa37811776f2","Type":"ContainerDied","Data":"6c087c6362dbae6d07d6f265a00659556f6a673ca5628661b42cc4827d724301"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.728911 4675 generic.go:334] "Generic (PLEG): container finished" podID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerID="440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7" exitCode=0 Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.729085 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7b6d" event={"ID":"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4","Type":"ContainerDied","Data":"440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.729126 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7b6d" event={"ID":"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4","Type":"ContainerStarted","Data":"cc79f8b191ea0f8e74a3eefe4ff2f89e27322abc8c06e9a72d85257c176a3ab8"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.746232 4675 generic.go:334] "Generic (PLEG): container finished" podID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerID="60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3" exitCode=0 Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.746483 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-catalog-content\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.746877 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-utilities\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.746952 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7ch8\" (UniqueName: \"kubernetes.io/projected/b2ff348d-9b80-4358-afcf-dace55aef217-kube-api-access-p7ch8\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.747303 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc6hh" event={"ID":"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2","Type":"ContainerDied","Data":"60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3"} Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.820547 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" podStartSLOduration=131.820528226 podStartE2EDuration="2m11.820528226s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:10.76911034 +0000 UTC m=+152.396200628" watchObservedRunningTime="2026-02-19 08:46:10.820528226 +0000 UTC m=+152.447618494" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.848476 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-catalog-content\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.848623 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-utilities\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.848688 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7ch8\" (UniqueName: \"kubernetes.io/projected/b2ff348d-9b80-4358-afcf-dace55aef217-kube-api-access-p7ch8\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.849804 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-utilities\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.850254 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-catalog-content\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.875304 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7ch8\" (UniqueName: \"kubernetes.io/projected/b2ff348d-9b80-4358-afcf-dace55aef217-kube-api-access-p7ch8\") pod \"redhat-marketplace-sgxzf\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:10 crc kubenswrapper[4675]: I0219 08:46:10.957369 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.003216 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n5xqw"] Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.004374 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.026518 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n5xqw"] Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.116078 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.152843 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq8sf\" (UniqueName: \"kubernetes.io/projected/57d4e815-2507-4852-b5be-40876f730e74-kube-api-access-wq8sf\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.152885 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-utilities\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.152944 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-catalog-content\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.160953 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.254416 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kubelet-dir\") pod \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.254575 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kube-api-access\") pod \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\" (UID: \"1bbe28ee-f09f-45c1-aae3-fa37811776f2\") " Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.254812 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq8sf\" (UniqueName: \"kubernetes.io/projected/57d4e815-2507-4852-b5be-40876f730e74-kube-api-access-wq8sf\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.254842 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-utilities\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.254903 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-catalog-content\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.256017 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-catalog-content\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.256324 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-utilities\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.256320 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1bbe28ee-f09f-45c1-aae3-fa37811776f2" (UID: "1bbe28ee-f09f-45c1-aae3-fa37811776f2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.262866 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1bbe28ee-f09f-45c1-aae3-fa37811776f2" (UID: "1bbe28ee-f09f-45c1-aae3-fa37811776f2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.281977 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq8sf\" (UniqueName: \"kubernetes.io/projected/57d4e815-2507-4852-b5be-40876f730e74-kube-api-access-wq8sf\") pod \"redhat-marketplace-n5xqw\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.324512 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgxzf"] Feb 19 08:46:11 crc kubenswrapper[4675]: W0219 08:46:11.341008 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2ff348d_9b80_4358_afcf_dace55aef217.slice/crio-1aec6ff40791d740027d8dce7864b4dd4b344a7fd1ed543160457ca5e0b32db8 WatchSource:0}: Error finding container 1aec6ff40791d740027d8dce7864b4dd4b344a7fd1ed543160457ca5e0b32db8: Status 404 returned error can't find the container with id 1aec6ff40791d740027d8dce7864b4dd4b344a7fd1ed543160457ca5e0b32db8 Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.341244 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.356770 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.356802 4675 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bbe28ee-f09f-45c1-aae3-fa37811776f2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.609271 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6mmqj"] Feb 19 08:46:11 crc kubenswrapper[4675]: E0219 08:46:11.610072 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bbe28ee-f09f-45c1-aae3-fa37811776f2" containerName="pruner" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.610090 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bbe28ee-f09f-45c1-aae3-fa37811776f2" containerName="pruner" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.610221 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bbe28ee-f09f-45c1-aae3-fa37811776f2" containerName="pruner" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.611157 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.614927 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.623999 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mmqj"] Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.640980 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.646030 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:11 crc kubenswrapper[4675]: [-]has-synced failed: reason withheld Feb 19 08:46:11 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:11 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.646085 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.670592 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n5xqw"] Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.721292 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.744184 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.744254 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.753949 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgxzf" event={"ID":"b2ff348d-9b80-4358-afcf-dace55aef217","Type":"ContainerStarted","Data":"1aec6ff40791d740027d8dce7864b4dd4b344a7fd1ed543160457ca5e0b32db8"} Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.757274 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1bbe28ee-f09f-45c1-aae3-fa37811776f2","Type":"ContainerDied","Data":"99428a340e7d29a16eaf648b3e60eddb411611bad1a0c9bcd7502d61412869a6"} Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.757325 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99428a340e7d29a16eaf648b3e60eddb411611bad1a0c9bcd7502d61412869a6" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.757400 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.767725 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n5xqw" event={"ID":"57d4e815-2507-4852-b5be-40876f730e74","Type":"ContainerStarted","Data":"f4462b9c91492b90ced0bfb8a9e8c88e0d93707dff2a088fb9b810e6ef3cd0ee"} Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.770604 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlhr2\" (UniqueName: \"kubernetes.io/projected/60ff6ed9-e694-4050-aae5-f22f390208e7-kube-api-access-jlhr2\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.770845 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-catalog-content\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.770907 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-utilities\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.872709 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-catalog-content\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.873173 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-utilities\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.873247 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlhr2\" (UniqueName: \"kubernetes.io/projected/60ff6ed9-e694-4050-aae5-f22f390208e7-kube-api-access-jlhr2\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.875620 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-utilities\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.875942 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-catalog-content\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.900572 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlhr2\" (UniqueName: \"kubernetes.io/projected/60ff6ed9-e694-4050-aae5-f22f390208e7-kube-api-access-jlhr2\") pod \"redhat-operators-6mmqj\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:11 crc kubenswrapper[4675]: I0219 08:46:11.942112 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.007269 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gr52t"] Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.008572 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.034722 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gr52t"] Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.110456 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.178092 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-catalog-content\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.178181 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjs5g\" (UniqueName: \"kubernetes.io/projected/271b0494-814e-4897-9f09-d2f6b75fe66e-kube-api-access-sjs5g\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.178219 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-utilities\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.223160 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mmqj"] Feb 19 08:46:12 crc kubenswrapper[4675]: W0219 08:46:12.230498 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60ff6ed9_e694_4050_aae5_f22f390208e7.slice/crio-02486411a3a4ea0a7e60ca4a5774e94f083421cb95abdabafd20589fc0799d38 WatchSource:0}: Error finding container 02486411a3a4ea0a7e60ca4a5774e94f083421cb95abdabafd20589fc0799d38: Status 404 returned error can't find the container with id 02486411a3a4ea0a7e60ca4a5774e94f083421cb95abdabafd20589fc0799d38 Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.279387 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-config-volume\") pod \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.279900 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-secret-volume\") pod \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.280013 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqwsv\" (UniqueName: \"kubernetes.io/projected/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-kube-api-access-bqwsv\") pod \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\" (UID: \"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd\") " Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.280159 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjs5g\" (UniqueName: \"kubernetes.io/projected/271b0494-814e-4897-9f09-d2f6b75fe66e-kube-api-access-sjs5g\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.280196 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-utilities\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.280262 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-catalog-content\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.280593 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-config-volume" (OuterVolumeSpecName: "config-volume") pod "e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" (UID: "e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.280753 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-catalog-content\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.281386 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-utilities\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.287632 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" (UID: "e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.287831 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-kube-api-access-bqwsv" (OuterVolumeSpecName: "kube-api-access-bqwsv") pod "e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" (UID: "e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd"). InnerVolumeSpecName "kube-api-access-bqwsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.305446 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjs5g\" (UniqueName: \"kubernetes.io/projected/271b0494-814e-4897-9f09-d2f6b75fe66e-kube-api-access-sjs5g\") pod \"redhat-operators-gr52t\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.341053 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.346975 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 19 08:46:12 crc kubenswrapper[4675]: E0219 08:46:12.347549 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" containerName="collect-profiles" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.347622 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" containerName="collect-profiles" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.347837 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" containerName="collect-profiles" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.348330 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.350664 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.351004 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.361898 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.390279 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqwsv\" (UniqueName: \"kubernetes.io/projected/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-kube-api-access-bqwsv\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.390315 4675 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-config-volume\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.390325 4675 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.493689 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.493746 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.569241 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gr52t"] Feb 19 08:46:12 crc kubenswrapper[4675]: W0219 08:46:12.580041 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod271b0494_814e_4897_9f09_d2f6b75fe66e.slice/crio-17b4b03edb8fbe9ad7ebc7cada2fe02b586aae56a8505fad1b18c805f3dbe89f WatchSource:0}: Error finding container 17b4b03edb8fbe9ad7ebc7cada2fe02b586aae56a8505fad1b18c805f3dbe89f: Status 404 returned error can't find the container with id 17b4b03edb8fbe9ad7ebc7cada2fe02b586aae56a8505fad1b18c805f3dbe89f Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.598155 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.598225 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.598337 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.618748 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.646110 4675 patch_prober.go:28] interesting pod/router-default-5444994796-rjqzp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 19 08:46:12 crc kubenswrapper[4675]: [+]has-synced ok Feb 19 08:46:12 crc kubenswrapper[4675]: [+]process-running ok Feb 19 08:46:12 crc kubenswrapper[4675]: healthz check failed Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.646203 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rjqzp" podUID="d3d57965-61a8-4a6f-883b-20360fd36509" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.718651 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.776539 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gr52t" event={"ID":"271b0494-814e-4897-9f09-d2f6b75fe66e","Type":"ContainerStarted","Data":"17b4b03edb8fbe9ad7ebc7cada2fe02b586aae56a8505fad1b18c805f3dbe89f"} Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.781592 4675 generic.go:334] "Generic (PLEG): container finished" podID="b2ff348d-9b80-4358-afcf-dace55aef217" containerID="9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250" exitCode=0 Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.781714 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgxzf" event={"ID":"b2ff348d-9b80-4358-afcf-dace55aef217","Type":"ContainerDied","Data":"9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250"} Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.790162 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.790186 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n" event={"ID":"e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd","Type":"ContainerDied","Data":"1880c29a43d8fae6c5cf8345bdb74b87bae1f13d36ff7b236fb1ac9ed332f789"} Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.790235 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1880c29a43d8fae6c5cf8345bdb74b87bae1f13d36ff7b236fb1ac9ed332f789" Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.796444 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerStarted","Data":"8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d"} Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.796507 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerStarted","Data":"02486411a3a4ea0a7e60ca4a5774e94f083421cb95abdabafd20589fc0799d38"} Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.811713 4675 generic.go:334] "Generic (PLEG): container finished" podID="57d4e815-2507-4852-b5be-40876f730e74" containerID="9c83ef2f6087b4cd5b05d389f1951d61dcf4faf2fbbf8fb2bafc109bb8dad39b" exitCode=0 Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.812940 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n5xqw" event={"ID":"57d4e815-2507-4852-b5be-40876f730e74","Type":"ContainerDied","Data":"9c83ef2f6087b4cd5b05d389f1951d61dcf4faf2fbbf8fb2bafc109bb8dad39b"} Feb 19 08:46:12 crc kubenswrapper[4675]: I0219 08:46:12.948867 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.646406 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.649584 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-rjqzp" Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.843164 4675 generic.go:334] "Generic (PLEG): container finished" podID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerID="d44eb74b19ba4bb35ca7171e9f184ba2d09e6ab4b3e77d309ac9732507e00e00" exitCode=0 Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.843341 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gr52t" event={"ID":"271b0494-814e-4897-9f09-d2f6b75fe66e","Type":"ContainerDied","Data":"d44eb74b19ba4bb35ca7171e9f184ba2d09e6ab4b3e77d309ac9732507e00e00"} Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.849691 4675 generic.go:334] "Generic (PLEG): container finished" podID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerID="8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d" exitCode=0 Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.849864 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerDied","Data":"8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d"} Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.861310 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2","Type":"ContainerStarted","Data":"5aff7a0deb50d8a1a162f4cbf3368a306f7b0538fd771f01b99163a692559a98"} Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.861354 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2","Type":"ContainerStarted","Data":"fe5c8d02babf1d5272cd74613b9459f47e375c621d94c7bc320ce215acf7e672"} Feb 19 08:46:13 crc kubenswrapper[4675]: I0219 08:46:13.912247 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.9122174730000001 podStartE2EDuration="1.912217473s" podCreationTimestamp="2026-02-19 08:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:13.887162171 +0000 UTC m=+155.514252439" watchObservedRunningTime="2026-02-19 08:46:13.912217473 +0000 UTC m=+155.539307741" Feb 19 08:46:14 crc kubenswrapper[4675]: I0219 08:46:14.880910 4675 generic.go:334] "Generic (PLEG): container finished" podID="300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2" containerID="5aff7a0deb50d8a1a162f4cbf3368a306f7b0538fd771f01b99163a692559a98" exitCode=0 Feb 19 08:46:14 crc kubenswrapper[4675]: I0219 08:46:14.880966 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2","Type":"ContainerDied","Data":"5aff7a0deb50d8a1a162f4cbf3368a306f7b0538fd771f01b99163a692559a98"} Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.290907 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.417234 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kubelet-dir\") pod \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.417416 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kube-api-access\") pod \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\" (UID: \"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2\") " Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.417749 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2" (UID: "300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.440371 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2" (UID: "300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.519273 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.519751 4675 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.813213 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6v99v" Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.929015 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2","Type":"ContainerDied","Data":"fe5c8d02babf1d5272cd74613b9459f47e375c621d94c7bc320ce215acf7e672"} Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.929139 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe5c8d02babf1d5272cd74613b9459f47e375c621d94c7bc320ce215acf7e672" Feb 19 08:46:16 crc kubenswrapper[4675]: I0219 08:46:16.929069 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 19 08:46:20 crc kubenswrapper[4675]: I0219 08:46:20.134882 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-pr5cf" Feb 19 08:46:20 crc kubenswrapper[4675]: I0219 08:46:20.587195 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:46:20 crc kubenswrapper[4675]: I0219 08:46:20.592580 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:46:22 crc kubenswrapper[4675]: I0219 08:46:22.144075 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:46:22 crc kubenswrapper[4675]: I0219 08:46:22.172767 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/58bc9d86-8ee5-4635-9de3-23890597f5ca-metrics-certs\") pod \"network-metrics-daemon-vcqbz\" (UID: \"58bc9d86-8ee5-4635-9de3-23890597f5ca\") " pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:46:22 crc kubenswrapper[4675]: I0219 08:46:22.439376 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vcqbz" Feb 19 08:46:29 crc kubenswrapper[4675]: I0219 08:46:29.693961 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:46:41 crc kubenswrapper[4675]: I0219 08:46:41.743895 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:46:41 crc kubenswrapper[4675]: I0219 08:46:41.744798 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:46:41 crc kubenswrapper[4675]: I0219 08:46:41.755426 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wt2kk" Feb 19 08:46:46 crc kubenswrapper[4675]: I0219 08:46:46.728054 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 19 08:46:46 crc kubenswrapper[4675]: E0219 08:46:46.867958 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 19 08:46:46 crc kubenswrapper[4675]: E0219 08:46:46.868162 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jlhr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6mmqj_openshift-marketplace(60ff6ed9-e694-4050-aae5-f22f390208e7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 19 08:46:46 crc kubenswrapper[4675]: E0219 08:46:46.869419 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-6mmqj" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" Feb 19 08:46:47 crc kubenswrapper[4675]: I0219 08:46:47.949662 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 19 08:46:47 crc kubenswrapper[4675]: E0219 08:46:47.950221 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2" containerName="pruner" Feb 19 08:46:47 crc kubenswrapper[4675]: I0219 08:46:47.950234 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2" containerName="pruner" Feb 19 08:46:47 crc kubenswrapper[4675]: I0219 08:46:47.950322 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="300c50dc-1c2d-47ba-b5ac-72a9bc19e5e2" containerName="pruner" Feb 19 08:46:47 crc kubenswrapper[4675]: I0219 08:46:47.950771 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:47 crc kubenswrapper[4675]: I0219 08:46:47.953445 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 19 08:46:47 crc kubenswrapper[4675]: I0219 08:46:47.953542 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 19 08:46:47 crc kubenswrapper[4675]: I0219 08:46:47.965462 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.034608 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83117ad9-9ce3-4a1f-a424-44b087e05c39-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.034672 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83117ad9-9ce3-4a1f-a424-44b087e05c39-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.137721 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83117ad9-9ce3-4a1f-a424-44b087e05c39-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.137771 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83117ad9-9ce3-4a1f-a424-44b087e05c39-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.138010 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83117ad9-9ce3-4a1f-a424-44b087e05c39-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.157532 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83117ad9-9ce3-4a1f-a424-44b087e05c39-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.303418 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.324936 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6mmqj" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.415574 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.415828 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzjmh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-t7b6d_openshift-marketplace(7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.417239 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-t7b6d" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.465089 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.465748 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sjs5g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gr52t_openshift-marketplace(271b0494-814e-4897-9f09-d2f6b75fe66e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.467014 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gr52t" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.496428 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.496605 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pskkw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vc6hh_openshift-marketplace(cfcc9c58-9792-449b-9fb8-faa5ef40cbc2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 19 08:46:48 crc kubenswrapper[4675]: E0219 08:46:48.498104 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vc6hh" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.644117 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 19 08:46:48 crc kubenswrapper[4675]: I0219 08:46:48.795792 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vcqbz"] Feb 19 08:46:48 crc kubenswrapper[4675]: W0219 08:46:48.810257 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58bc9d86_8ee5_4635_9de3_23890597f5ca.slice/crio-4bfd9cb4f6fa21c183e95ae0c732b6d3cbe8c1a2c2ea2a9c35624e476478a545 WatchSource:0}: Error finding container 4bfd9cb4f6fa21c183e95ae0c732b6d3cbe8c1a2c2ea2a9c35624e476478a545: Status 404 returned error can't find the container with id 4bfd9cb4f6fa21c183e95ae0c732b6d3cbe8c1a2c2ea2a9c35624e476478a545 Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.154092 4675 generic.go:334] "Generic (PLEG): container finished" podID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerID="47e856338d65d76a9f27a9ca04dde940d0fc5811bf691561a9a986e3417f8888" exitCode=0 Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.154179 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kk7x5" event={"ID":"fd1115a1-2835-4bc5-8d17-7387b7f34cc4","Type":"ContainerDied","Data":"47e856338d65d76a9f27a9ca04dde940d0fc5811bf691561a9a986e3417f8888"} Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.158167 4675 generic.go:334] "Generic (PLEG): container finished" podID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerID="b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5" exitCode=0 Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.158260 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbjkd" event={"ID":"52c2e6ec-6246-4346-8d44-f9d37b805bb0","Type":"ContainerDied","Data":"b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5"} Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.161066 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" event={"ID":"58bc9d86-8ee5-4635-9de3-23890597f5ca","Type":"ContainerStarted","Data":"4bfd9cb4f6fa21c183e95ae0c732b6d3cbe8c1a2c2ea2a9c35624e476478a545"} Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.164200 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83117ad9-9ce3-4a1f-a424-44b087e05c39","Type":"ContainerStarted","Data":"040ff5b6ebdae566e4960c1131fd7362065c80545c02c2624d0fec4db4dfe0fd"} Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.164326 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83117ad9-9ce3-4a1f-a424-44b087e05c39","Type":"ContainerStarted","Data":"7bc661259d8cc2c38f73fb70235fada1c3b6f6389eb1aba71ff25224283cc19c"} Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.166959 4675 generic.go:334] "Generic (PLEG): container finished" podID="57d4e815-2507-4852-b5be-40876f730e74" containerID="a689a266738347e717cf66fa3cfddd00fdad31a15a4443b57a6ab45de7ac732d" exitCode=0 Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.167081 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n5xqw" event={"ID":"57d4e815-2507-4852-b5be-40876f730e74","Type":"ContainerDied","Data":"a689a266738347e717cf66fa3cfddd00fdad31a15a4443b57a6ab45de7ac732d"} Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.175726 4675 generic.go:334] "Generic (PLEG): container finished" podID="b2ff348d-9b80-4358-afcf-dace55aef217" containerID="79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837" exitCode=0 Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.177520 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgxzf" event={"ID":"b2ff348d-9b80-4358-afcf-dace55aef217","Type":"ContainerDied","Data":"79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837"} Feb 19 08:46:49 crc kubenswrapper[4675]: E0219 08:46:49.178231 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-t7b6d" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" Feb 19 08:46:49 crc kubenswrapper[4675]: E0219 08:46:49.179609 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gr52t" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" Feb 19 08:46:49 crc kubenswrapper[4675]: E0219 08:46:49.182039 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vc6hh" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" Feb 19 08:46:49 crc kubenswrapper[4675]: I0219 08:46:49.214167 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.214144222 podStartE2EDuration="2.214144222s" podCreationTimestamp="2026-02-19 08:46:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:49.209196601 +0000 UTC m=+190.836286879" watchObservedRunningTime="2026-02-19 08:46:49.214144222 +0000 UTC m=+190.841234500" Feb 19 08:46:50 crc kubenswrapper[4675]: I0219 08:46:50.190209 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" event={"ID":"58bc9d86-8ee5-4635-9de3-23890597f5ca","Type":"ContainerStarted","Data":"2fc3e84ab6c67b48a629c20a9d97d33e493128068a6e9bda783794cf50c8dc69"} Feb 19 08:46:50 crc kubenswrapper[4675]: I0219 08:46:50.190650 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vcqbz" event={"ID":"58bc9d86-8ee5-4635-9de3-23890597f5ca","Type":"ContainerStarted","Data":"89fda915f964ba0d6acc8241b17f75d4fe1b442b1cddd2e2b9731fe04af8c788"} Feb 19 08:46:50 crc kubenswrapper[4675]: I0219 08:46:50.199272 4675 generic.go:334] "Generic (PLEG): container finished" podID="83117ad9-9ce3-4a1f-a424-44b087e05c39" containerID="040ff5b6ebdae566e4960c1131fd7362065c80545c02c2624d0fec4db4dfe0fd" exitCode=0 Feb 19 08:46:50 crc kubenswrapper[4675]: I0219 08:46:50.199327 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83117ad9-9ce3-4a1f-a424-44b087e05c39","Type":"ContainerDied","Data":"040ff5b6ebdae566e4960c1131fd7362065c80545c02c2624d0fec4db4dfe0fd"} Feb 19 08:46:50 crc kubenswrapper[4675]: I0219 08:46:50.213382 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vcqbz" podStartSLOduration=171.213357719 podStartE2EDuration="2m51.213357719s" podCreationTimestamp="2026-02-19 08:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:50.212892024 +0000 UTC m=+191.839982312" watchObservedRunningTime="2026-02-19 08:46:50.213357719 +0000 UTC m=+191.840447997" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.210452 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgxzf" event={"ID":"b2ff348d-9b80-4358-afcf-dace55aef217","Type":"ContainerStarted","Data":"66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd"} Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.214485 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kk7x5" event={"ID":"fd1115a1-2835-4bc5-8d17-7387b7f34cc4","Type":"ContainerStarted","Data":"27936572cd7b4953e5dd2b409a82652e57b135d841c6f4f1779c14470e1d7ecb"} Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.218997 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbjkd" event={"ID":"52c2e6ec-6246-4346-8d44-f9d37b805bb0","Type":"ContainerStarted","Data":"7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700"} Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.222834 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n5xqw" event={"ID":"57d4e815-2507-4852-b5be-40876f730e74","Type":"ContainerStarted","Data":"b472dcbc9a660ddbf76e0223612997382bb2bd9ee8c20ee85570d0a8cb50b737"} Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.241986 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sgxzf" podStartSLOduration=3.525511314 podStartE2EDuration="41.241845905s" podCreationTimestamp="2026-02-19 08:46:10 +0000 UTC" firstStartedPulling="2026-02-19 08:46:12.785516846 +0000 UTC m=+154.412607114" lastFinishedPulling="2026-02-19 08:46:50.501851437 +0000 UTC m=+192.128941705" observedRunningTime="2026-02-19 08:46:51.240577063 +0000 UTC m=+192.867667341" watchObservedRunningTime="2026-02-19 08:46:51.241845905 +0000 UTC m=+192.868936173" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.291220 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n5xqw" podStartSLOduration=3.855360872 podStartE2EDuration="41.291194664s" podCreationTimestamp="2026-02-19 08:46:10 +0000 UTC" firstStartedPulling="2026-02-19 08:46:12.818421882 +0000 UTC m=+154.445512140" lastFinishedPulling="2026-02-19 08:46:50.254255664 +0000 UTC m=+191.881345932" observedRunningTime="2026-02-19 08:46:51.266152362 +0000 UTC m=+192.893242640" watchObservedRunningTime="2026-02-19 08:46:51.291194664 +0000 UTC m=+192.918284932" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.293371 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kk7x5" podStartSLOduration=3.626067498 podStartE2EDuration="43.293357724s" podCreationTimestamp="2026-02-19 08:46:08 +0000 UTC" firstStartedPulling="2026-02-19 08:46:10.65307544 +0000 UTC m=+152.280165708" lastFinishedPulling="2026-02-19 08:46:50.320365666 +0000 UTC m=+191.947455934" observedRunningTime="2026-02-19 08:46:51.290377297 +0000 UTC m=+192.917467565" watchObservedRunningTime="2026-02-19 08:46:51.293357724 +0000 UTC m=+192.920447992" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.314137 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vbjkd" podStartSLOduration=3.840348402 podStartE2EDuration="43.314108536s" podCreationTimestamp="2026-02-19 08:46:08 +0000 UTC" firstStartedPulling="2026-02-19 08:46:10.669252565 +0000 UTC m=+152.296342833" lastFinishedPulling="2026-02-19 08:46:50.143012699 +0000 UTC m=+191.770102967" observedRunningTime="2026-02-19 08:46:51.309722314 +0000 UTC m=+192.936812582" watchObservedRunningTime="2026-02-19 08:46:51.314108536 +0000 UTC m=+192.941198804" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.342493 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.342576 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.568162 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.609303 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83117ad9-9ce3-4a1f-a424-44b087e05c39-kube-api-access\") pod \"83117ad9-9ce3-4a1f-a424-44b087e05c39\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.609429 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83117ad9-9ce3-4a1f-a424-44b087e05c39-kubelet-dir\") pod \"83117ad9-9ce3-4a1f-a424-44b087e05c39\" (UID: \"83117ad9-9ce3-4a1f-a424-44b087e05c39\") " Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.609842 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83117ad9-9ce3-4a1f-a424-44b087e05c39-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "83117ad9-9ce3-4a1f-a424-44b087e05c39" (UID: "83117ad9-9ce3-4a1f-a424-44b087e05c39"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.621924 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83117ad9-9ce3-4a1f-a424-44b087e05c39-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "83117ad9-9ce3-4a1f-a424-44b087e05c39" (UID: "83117ad9-9ce3-4a1f-a424-44b087e05c39"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.711593 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83117ad9-9ce3-4a1f-a424-44b087e05c39-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:51 crc kubenswrapper[4675]: I0219 08:46:51.711665 4675 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83117ad9-9ce3-4a1f-a424-44b087e05c39-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.230056 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83117ad9-9ce3-4a1f-a424-44b087e05c39","Type":"ContainerDied","Data":"7bc661259d8cc2c38f73fb70235fada1c3b6f6389eb1aba71ff25224283cc19c"} Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.230539 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bc661259d8cc2c38f73fb70235fada1c3b6f6389eb1aba71ff25224283cc19c" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.230172 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.546253 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-n5xqw" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="registry-server" probeResult="failure" output=< Feb 19 08:46:52 crc kubenswrapper[4675]: timeout: failed to connect service ":50051" within 1s Feb 19 08:46:52 crc kubenswrapper[4675]: > Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.945923 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 19 08:46:52 crc kubenswrapper[4675]: E0219 08:46:52.946221 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83117ad9-9ce3-4a1f-a424-44b087e05c39" containerName="pruner" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.946239 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="83117ad9-9ce3-4a1f-a424-44b087e05c39" containerName="pruner" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.946360 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="83117ad9-9ce3-4a1f-a424-44b087e05c39" containerName="pruner" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.946839 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.950056 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.950941 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 19 08:46:52 crc kubenswrapper[4675]: I0219 08:46:52.955467 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.037197 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78e84b11-6e92-4596-9397-017767524db7-kube-api-access\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.037271 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.037354 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-var-lock\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.138512 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78e84b11-6e92-4596-9397-017767524db7-kube-api-access\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.138905 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.138622 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-kubelet-dir\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.139089 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-var-lock\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.139324 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-var-lock\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.176880 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78e84b11-6e92-4596-9397-017767524db7-kube-api-access\") pod \"installer-9-crc\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.273341 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:46:53 crc kubenswrapper[4675]: I0219 08:46:53.536706 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 19 08:46:53 crc kubenswrapper[4675]: W0219 08:46:53.553245 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod78e84b11_6e92_4596_9397_017767524db7.slice/crio-65eb77593ef4d467e6644496d5fa6015089e4e45a65cac7a3a8b4b363dbc87eb WatchSource:0}: Error finding container 65eb77593ef4d467e6644496d5fa6015089e4e45a65cac7a3a8b4b363dbc87eb: Status 404 returned error can't find the container with id 65eb77593ef4d467e6644496d5fa6015089e4e45a65cac7a3a8b4b363dbc87eb Feb 19 08:46:54 crc kubenswrapper[4675]: I0219 08:46:54.243859 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"78e84b11-6e92-4596-9397-017767524db7","Type":"ContainerStarted","Data":"65eb77593ef4d467e6644496d5fa6015089e4e45a65cac7a3a8b4b363dbc87eb"} Feb 19 08:46:55 crc kubenswrapper[4675]: I0219 08:46:55.251306 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"78e84b11-6e92-4596-9397-017767524db7","Type":"ContainerStarted","Data":"e7a94ddb33287eccfc95bb313522f7d43d5206d8127b7d83b8f9a70154c44ed7"} Feb 19 08:46:55 crc kubenswrapper[4675]: I0219 08:46:55.269602 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.269561713 podStartE2EDuration="3.269561713s" podCreationTimestamp="2026-02-19 08:46:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:46:55.268113966 +0000 UTC m=+196.895204234" watchObservedRunningTime="2026-02-19 08:46:55.269561713 +0000 UTC m=+196.896651981" Feb 19 08:46:58 crc kubenswrapper[4675]: I0219 08:46:58.987208 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:58 crc kubenswrapper[4675]: I0219 08:46:58.987597 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:59 crc kubenswrapper[4675]: I0219 08:46:59.048399 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:59 crc kubenswrapper[4675]: I0219 08:46:59.315123 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:46:59 crc kubenswrapper[4675]: I0219 08:46:59.638730 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:59 crc kubenswrapper[4675]: I0219 08:46:59.638830 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:46:59 crc kubenswrapper[4675]: I0219 08:46:59.659525 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-plmm7"] Feb 19 08:46:59 crc kubenswrapper[4675]: I0219 08:46:59.729945 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:47:00 crc kubenswrapper[4675]: I0219 08:47:00.324603 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:47:00 crc kubenswrapper[4675]: I0219 08:47:00.958604 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:47:00 crc kubenswrapper[4675]: I0219 08:47:00.958972 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:47:01 crc kubenswrapper[4675]: I0219 08:47:01.001208 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:47:01 crc kubenswrapper[4675]: I0219 08:47:01.327841 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:47:01 crc kubenswrapper[4675]: I0219 08:47:01.386071 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:47:01 crc kubenswrapper[4675]: I0219 08:47:01.428095 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:47:02 crc kubenswrapper[4675]: I0219 08:47:02.683047 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kk7x5"] Feb 19 08:47:02 crc kubenswrapper[4675]: I0219 08:47:02.683399 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kk7x5" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="registry-server" containerID="cri-o://27936572cd7b4953e5dd2b409a82652e57b135d841c6f4f1779c14470e1d7ecb" gracePeriod=2 Feb 19 08:47:03 crc kubenswrapper[4675]: I0219 08:47:03.302319 4675 generic.go:334] "Generic (PLEG): container finished" podID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerID="27936572cd7b4953e5dd2b409a82652e57b135d841c6f4f1779c14470e1d7ecb" exitCode=0 Feb 19 08:47:03 crc kubenswrapper[4675]: I0219 08:47:03.302812 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kk7x5" event={"ID":"fd1115a1-2835-4bc5-8d17-7387b7f34cc4","Type":"ContainerDied","Data":"27936572cd7b4953e5dd2b409a82652e57b135d841c6f4f1779c14470e1d7ecb"} Feb 19 08:47:04 crc kubenswrapper[4675]: I0219 08:47:04.081374 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n5xqw"] Feb 19 08:47:04 crc kubenswrapper[4675]: I0219 08:47:04.082392 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n5xqw" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="registry-server" containerID="cri-o://b472dcbc9a660ddbf76e0223612997382bb2bd9ee8c20ee85570d0a8cb50b737" gracePeriod=2 Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.335049 4675 generic.go:334] "Generic (PLEG): container finished" podID="57d4e815-2507-4852-b5be-40876f730e74" containerID="b472dcbc9a660ddbf76e0223612997382bb2bd9ee8c20ee85570d0a8cb50b737" exitCode=0 Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.335379 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n5xqw" event={"ID":"57d4e815-2507-4852-b5be-40876f730e74","Type":"ContainerDied","Data":"b472dcbc9a660ddbf76e0223612997382bb2bd9ee8c20ee85570d0a8cb50b737"} Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.516230 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.628775 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm4j6\" (UniqueName: \"kubernetes.io/projected/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-kube-api-access-gm4j6\") pod \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.628823 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-catalog-content\") pod \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.628981 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-utilities\") pod \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\" (UID: \"fd1115a1-2835-4bc5-8d17-7387b7f34cc4\") " Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.629981 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-utilities" (OuterVolumeSpecName: "utilities") pod "fd1115a1-2835-4bc5-8d17-7387b7f34cc4" (UID: "fd1115a1-2835-4bc5-8d17-7387b7f34cc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.635439 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-kube-api-access-gm4j6" (OuterVolumeSpecName: "kube-api-access-gm4j6") pod "fd1115a1-2835-4bc5-8d17-7387b7f34cc4" (UID: "fd1115a1-2835-4bc5-8d17-7387b7f34cc4"). InnerVolumeSpecName "kube-api-access-gm4j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.656752 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.686030 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd1115a1-2835-4bc5-8d17-7387b7f34cc4" (UID: "fd1115a1-2835-4bc5-8d17-7387b7f34cc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.730162 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-utilities\") pod \"57d4e815-2507-4852-b5be-40876f730e74\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.730301 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq8sf\" (UniqueName: \"kubernetes.io/projected/57d4e815-2507-4852-b5be-40876f730e74-kube-api-access-wq8sf\") pod \"57d4e815-2507-4852-b5be-40876f730e74\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.730389 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-catalog-content\") pod \"57d4e815-2507-4852-b5be-40876f730e74\" (UID: \"57d4e815-2507-4852-b5be-40876f730e74\") " Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.730718 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.730737 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.730750 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm4j6\" (UniqueName: \"kubernetes.io/projected/fd1115a1-2835-4bc5-8d17-7387b7f34cc4-kube-api-access-gm4j6\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.730951 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-utilities" (OuterVolumeSpecName: "utilities") pod "57d4e815-2507-4852-b5be-40876f730e74" (UID: "57d4e815-2507-4852-b5be-40876f730e74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.734049 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d4e815-2507-4852-b5be-40876f730e74-kube-api-access-wq8sf" (OuterVolumeSpecName: "kube-api-access-wq8sf") pod "57d4e815-2507-4852-b5be-40876f730e74" (UID: "57d4e815-2507-4852-b5be-40876f730e74"). InnerVolumeSpecName "kube-api-access-wq8sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.755961 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57d4e815-2507-4852-b5be-40876f730e74" (UID: "57d4e815-2507-4852-b5be-40876f730e74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.832407 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.832458 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq8sf\" (UniqueName: \"kubernetes.io/projected/57d4e815-2507-4852-b5be-40876f730e74-kube-api-access-wq8sf\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:05 crc kubenswrapper[4675]: I0219 08:47:05.832474 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d4e815-2507-4852-b5be-40876f730e74-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.344860 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n5xqw" event={"ID":"57d4e815-2507-4852-b5be-40876f730e74","Type":"ContainerDied","Data":"f4462b9c91492b90ced0bfb8a9e8c88e0d93707dff2a088fb9b810e6ef3cd0ee"} Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.344898 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n5xqw" Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.344932 4675 scope.go:117] "RemoveContainer" containerID="b472dcbc9a660ddbf76e0223612997382bb2bd9ee8c20ee85570d0a8cb50b737" Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.360771 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kk7x5" event={"ID":"fd1115a1-2835-4bc5-8d17-7387b7f34cc4","Type":"ContainerDied","Data":"7d6fef11f2d7c48beb46e0e844fb2a8d247db06d853fe11152f110ece52f7392"} Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.361094 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kk7x5" Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.390958 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n5xqw"] Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.395651 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n5xqw"] Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.407964 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kk7x5"] Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.410760 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kk7x5"] Feb 19 08:47:06 crc kubenswrapper[4675]: I0219 08:47:06.834925 4675 scope.go:117] "RemoveContainer" containerID="a689a266738347e717cf66fa3cfddd00fdad31a15a4443b57a6ab45de7ac732d" Feb 19 08:47:07 crc kubenswrapper[4675]: I0219 08:47:07.110724 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57d4e815-2507-4852-b5be-40876f730e74" path="/var/lib/kubelet/pods/57d4e815-2507-4852-b5be-40876f730e74/volumes" Feb 19 08:47:07 crc kubenswrapper[4675]: I0219 08:47:07.111443 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" path="/var/lib/kubelet/pods/fd1115a1-2835-4bc5-8d17-7387b7f34cc4/volumes" Feb 19 08:47:07 crc kubenswrapper[4675]: I0219 08:47:07.617995 4675 scope.go:117] "RemoveContainer" containerID="9c83ef2f6087b4cd5b05d389f1951d61dcf4faf2fbbf8fb2bafc109bb8dad39b" Feb 19 08:47:07 crc kubenswrapper[4675]: I0219 08:47:07.643778 4675 scope.go:117] "RemoveContainer" containerID="27936572cd7b4953e5dd2b409a82652e57b135d841c6f4f1779c14470e1d7ecb" Feb 19 08:47:07 crc kubenswrapper[4675]: I0219 08:47:07.668422 4675 scope.go:117] "RemoveContainer" containerID="47e856338d65d76a9f27a9ca04dde940d0fc5811bf691561a9a986e3417f8888" Feb 19 08:47:07 crc kubenswrapper[4675]: I0219 08:47:07.688794 4675 scope.go:117] "RemoveContainer" containerID="d25eaae249951e387b7e626a1280c9c77de5bc0131d6bf83aa18a6ccce8aaad1" Feb 19 08:47:08 crc kubenswrapper[4675]: I0219 08:47:08.380415 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerStarted","Data":"00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876"} Feb 19 08:47:08 crc kubenswrapper[4675]: I0219 08:47:08.385181 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gr52t" event={"ID":"271b0494-814e-4897-9f09-d2f6b75fe66e","Type":"ContainerStarted","Data":"ac891083f817882cf3f1f2eeabd221d187078d71799a1c2c5bc54a34f89b0249"} Feb 19 08:47:08 crc kubenswrapper[4675]: I0219 08:47:08.387499 4675 generic.go:334] "Generic (PLEG): container finished" podID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerID="80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e" exitCode=0 Feb 19 08:47:08 crc kubenswrapper[4675]: I0219 08:47:08.387558 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc6hh" event={"ID":"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2","Type":"ContainerDied","Data":"80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e"} Feb 19 08:47:08 crc kubenswrapper[4675]: I0219 08:47:08.399950 4675 generic.go:334] "Generic (PLEG): container finished" podID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerID="cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f" exitCode=0 Feb 19 08:47:08 crc kubenswrapper[4675]: I0219 08:47:08.400067 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7b6d" event={"ID":"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4","Type":"ContainerDied","Data":"cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f"} Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.409427 4675 generic.go:334] "Generic (PLEG): container finished" podID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerID="ac891083f817882cf3f1f2eeabd221d187078d71799a1c2c5bc54a34f89b0249" exitCode=0 Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.409496 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gr52t" event={"ID":"271b0494-814e-4897-9f09-d2f6b75fe66e","Type":"ContainerDied","Data":"ac891083f817882cf3f1f2eeabd221d187078d71799a1c2c5bc54a34f89b0249"} Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.414291 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7b6d" event={"ID":"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4","Type":"ContainerStarted","Data":"cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333"} Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.419788 4675 generic.go:334] "Generic (PLEG): container finished" podID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerID="00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876" exitCode=0 Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.419917 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerDied","Data":"00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876"} Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.423834 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc6hh" event={"ID":"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2","Type":"ContainerStarted","Data":"245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9"} Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.448869 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vc6hh" podStartSLOduration=3.419331483 podStartE2EDuration="1m1.44884181s" podCreationTimestamp="2026-02-19 08:46:08 +0000 UTC" firstStartedPulling="2026-02-19 08:46:10.748218073 +0000 UTC m=+152.375308341" lastFinishedPulling="2026-02-19 08:47:08.7777284 +0000 UTC m=+210.404818668" observedRunningTime="2026-02-19 08:47:09.447856919 +0000 UTC m=+211.074947197" watchObservedRunningTime="2026-02-19 08:47:09.44884181 +0000 UTC m=+211.075932078" Feb 19 08:47:09 crc kubenswrapper[4675]: I0219 08:47:09.491067 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t7b6d" podStartSLOduration=3.423962751 podStartE2EDuration="1m1.491044923s" podCreationTimestamp="2026-02-19 08:46:08 +0000 UTC" firstStartedPulling="2026-02-19 08:46:10.732580196 +0000 UTC m=+152.359670464" lastFinishedPulling="2026-02-19 08:47:08.799662378 +0000 UTC m=+210.426752636" observedRunningTime="2026-02-19 08:47:09.488128369 +0000 UTC m=+211.115218637" watchObservedRunningTime="2026-02-19 08:47:09.491044923 +0000 UTC m=+211.118135221" Feb 19 08:47:10 crc kubenswrapper[4675]: I0219 08:47:10.430680 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerStarted","Data":"59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6"} Feb 19 08:47:10 crc kubenswrapper[4675]: I0219 08:47:10.433178 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gr52t" event={"ID":"271b0494-814e-4897-9f09-d2f6b75fe66e","Type":"ContainerStarted","Data":"8b653f20181c4054b5443bc568c95ebf651f6a4eab883d26ae70d9fb60dfa383"} Feb 19 08:47:10 crc kubenswrapper[4675]: I0219 08:47:10.455865 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6mmqj" podStartSLOduration=3.486243073 podStartE2EDuration="59.455835077s" podCreationTimestamp="2026-02-19 08:46:11 +0000 UTC" firstStartedPulling="2026-02-19 08:46:13.854805653 +0000 UTC m=+155.481895921" lastFinishedPulling="2026-02-19 08:47:09.824397657 +0000 UTC m=+211.451487925" observedRunningTime="2026-02-19 08:47:10.453936726 +0000 UTC m=+212.081026994" watchObservedRunningTime="2026-02-19 08:47:10.455835077 +0000 UTC m=+212.082925345" Feb 19 08:47:11 crc kubenswrapper[4675]: I0219 08:47:11.743382 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:47:11 crc kubenswrapper[4675]: I0219 08:47:11.743451 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:47:11 crc kubenswrapper[4675]: I0219 08:47:11.743503 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:47:11 crc kubenswrapper[4675]: I0219 08:47:11.744283 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 08:47:11 crc kubenswrapper[4675]: I0219 08:47:11.744372 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e" gracePeriod=600 Feb 19 08:47:11 crc kubenswrapper[4675]: I0219 08:47:11.943725 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:47:11 crc kubenswrapper[4675]: I0219 08:47:11.944102 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:47:12 crc kubenswrapper[4675]: I0219 08:47:12.342297 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:47:12 crc kubenswrapper[4675]: I0219 08:47:12.342771 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:47:12 crc kubenswrapper[4675]: I0219 08:47:12.444177 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e" exitCode=0 Feb 19 08:47:12 crc kubenswrapper[4675]: I0219 08:47:12.444227 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e"} Feb 19 08:47:12 crc kubenswrapper[4675]: I0219 08:47:12.444259 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"ff7142b6a17f3a1cc3ed6d7db969da62d4d6cf49631898d5a6c5d25fdadb0c34"} Feb 19 08:47:12 crc kubenswrapper[4675]: I0219 08:47:12.460934 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gr52t" podStartSLOduration=5.424301632 podStartE2EDuration="1m1.460916341s" podCreationTimestamp="2026-02-19 08:46:11 +0000 UTC" firstStartedPulling="2026-02-19 08:46:13.852913441 +0000 UTC m=+155.480003709" lastFinishedPulling="2026-02-19 08:47:09.88952815 +0000 UTC m=+211.516618418" observedRunningTime="2026-02-19 08:47:10.48197189 +0000 UTC m=+212.109062158" watchObservedRunningTime="2026-02-19 08:47:12.460916341 +0000 UTC m=+214.088006609" Feb 19 08:47:12 crc kubenswrapper[4675]: I0219 08:47:12.995312 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6mmqj" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="registry-server" probeResult="failure" output=< Feb 19 08:47:12 crc kubenswrapper[4675]: timeout: failed to connect service ":50051" within 1s Feb 19 08:47:12 crc kubenswrapper[4675]: > Feb 19 08:47:13 crc kubenswrapper[4675]: I0219 08:47:13.384059 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gr52t" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="registry-server" probeResult="failure" output=< Feb 19 08:47:13 crc kubenswrapper[4675]: timeout: failed to connect service ":50051" within 1s Feb 19 08:47:13 crc kubenswrapper[4675]: > Feb 19 08:47:18 crc kubenswrapper[4675]: I0219 08:47:18.783833 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:47:18 crc kubenswrapper[4675]: I0219 08:47:18.784259 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:47:18 crc kubenswrapper[4675]: I0219 08:47:18.840254 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:47:19 crc kubenswrapper[4675]: I0219 08:47:19.205507 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:47:19 crc kubenswrapper[4675]: I0219 08:47:19.205950 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:47:19 crc kubenswrapper[4675]: I0219 08:47:19.246436 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:47:19 crc kubenswrapper[4675]: I0219 08:47:19.526590 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:47:19 crc kubenswrapper[4675]: I0219 08:47:19.529198 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:47:20 crc kubenswrapper[4675]: I0219 08:47:20.066535 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7b6d"] Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.497218 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t7b6d" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="registry-server" containerID="cri-o://cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333" gracePeriod=2 Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.837786 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.946568 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-utilities\") pod \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.946608 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzjmh\" (UniqueName: \"kubernetes.io/projected/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-kube-api-access-xzjmh\") pod \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.946715 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-catalog-content\") pod \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\" (UID: \"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4\") " Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.948033 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-utilities" (OuterVolumeSpecName: "utilities") pod "7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" (UID: "7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.954030 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-kube-api-access-xzjmh" (OuterVolumeSpecName: "kube-api-access-xzjmh") pod "7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" (UID: "7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4"). InnerVolumeSpecName "kube-api-access-xzjmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:47:21 crc kubenswrapper[4675]: I0219 08:47:21.986052 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.010656 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" (UID: "7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.023198 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.048286 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.048321 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzjmh\" (UniqueName: \"kubernetes.io/projected/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-kube-api-access-xzjmh\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.048331 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.383928 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.420865 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.504998 4675 generic.go:334] "Generic (PLEG): container finished" podID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerID="cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333" exitCode=0 Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.505096 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7b6d" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.505133 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7b6d" event={"ID":"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4","Type":"ContainerDied","Data":"cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333"} Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.505220 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7b6d" event={"ID":"7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4","Type":"ContainerDied","Data":"cc79f8b191ea0f8e74a3eefe4ff2f89e27322abc8c06e9a72d85257c176a3ab8"} Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.505260 4675 scope.go:117] "RemoveContainer" containerID="cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.528397 4675 scope.go:117] "RemoveContainer" containerID="cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.547764 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7b6d"] Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.551093 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t7b6d"] Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.557532 4675 scope.go:117] "RemoveContainer" containerID="440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.579625 4675 scope.go:117] "RemoveContainer" containerID="cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333" Feb 19 08:47:22 crc kubenswrapper[4675]: E0219 08:47:22.580117 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333\": container with ID starting with cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333 not found: ID does not exist" containerID="cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.580148 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333"} err="failed to get container status \"cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333\": rpc error: code = NotFound desc = could not find container \"cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333\": container with ID starting with cb79ac955c4620bc20c1298dfc0db30555b5cbeedc0a273c7a26ed4c150b1333 not found: ID does not exist" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.580170 4675 scope.go:117] "RemoveContainer" containerID="cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f" Feb 19 08:47:22 crc kubenswrapper[4675]: E0219 08:47:22.580443 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f\": container with ID starting with cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f not found: ID does not exist" containerID="cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.580465 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f"} err="failed to get container status \"cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f\": rpc error: code = NotFound desc = could not find container \"cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f\": container with ID starting with cc0c5b196dcbb38452f6dfa2bb9e13ac74b60ee2b654c710e37ff0bfe753a61f not found: ID does not exist" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.580480 4675 scope.go:117] "RemoveContainer" containerID="440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7" Feb 19 08:47:22 crc kubenswrapper[4675]: E0219 08:47:22.580849 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7\": container with ID starting with 440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7 not found: ID does not exist" containerID="440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7" Feb 19 08:47:22 crc kubenswrapper[4675]: I0219 08:47:22.580877 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7"} err="failed to get container status \"440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7\": rpc error: code = NotFound desc = could not find container \"440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7\": container with ID starting with 440e5f87fad570c90ed0ffe7553f1fa9039c6bb46c7f298678877f01ceb07ee7 not found: ID does not exist" Feb 19 08:47:23 crc kubenswrapper[4675]: I0219 08:47:23.110203 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" path="/var/lib/kubelet/pods/7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4/volumes" Feb 19 08:47:24 crc kubenswrapper[4675]: I0219 08:47:24.267085 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gr52t"] Feb 19 08:47:24 crc kubenswrapper[4675]: I0219 08:47:24.267342 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gr52t" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="registry-server" containerID="cri-o://8b653f20181c4054b5443bc568c95ebf651f6a4eab883d26ae70d9fb60dfa383" gracePeriod=2 Feb 19 08:47:24 crc kubenswrapper[4675]: I0219 08:47:24.692023 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" podUID="4d1c19bc-9031-49eb-88a1-3cb263134b3d" containerName="oauth-openshift" containerID="cri-o://de50c513a7f22497734c80a4bd667b0ad2672c40f7385eed2474943958cc320f" gracePeriod=15 Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.526048 4675 generic.go:334] "Generic (PLEG): container finished" podID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerID="8b653f20181c4054b5443bc568c95ebf651f6a4eab883d26ae70d9fb60dfa383" exitCode=0 Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.526092 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gr52t" event={"ID":"271b0494-814e-4897-9f09-d2f6b75fe66e","Type":"ContainerDied","Data":"8b653f20181c4054b5443bc568c95ebf651f6a4eab883d26ae70d9fb60dfa383"} Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.528300 4675 generic.go:334] "Generic (PLEG): container finished" podID="4d1c19bc-9031-49eb-88a1-3cb263134b3d" containerID="de50c513a7f22497734c80a4bd667b0ad2672c40f7385eed2474943958cc320f" exitCode=0 Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.528350 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" event={"ID":"4d1c19bc-9031-49eb-88a1-3cb263134b3d","Type":"ContainerDied","Data":"de50c513a7f22497734c80a4bd667b0ad2672c40f7385eed2474943958cc320f"} Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.776214 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.801814 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-router-certs\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.801889 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-provider-selection\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.801930 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-dir\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.801956 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-serving-cert\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802004 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-trusted-ca-bundle\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802048 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-session\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802079 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-policies\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802109 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-ocp-branding-template\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802139 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-cliconfig\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802166 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-service-ca\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802186 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-login\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802210 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbr8s\" (UniqueName: \"kubernetes.io/projected/4d1c19bc-9031-49eb-88a1-3cb263134b3d-kube-api-access-mbr8s\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802250 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-error\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.802287 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-idp-0-file-data\") pod \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\" (UID: \"4d1c19bc-9031-49eb-88a1-3cb263134b3d\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.804292 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.804623 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.804938 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.806514 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.808167 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.811569 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.814799 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.815983 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.816444 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.816930 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.817542 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.817829 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.818036 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.818412 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d1c19bc-9031-49eb-88a1-3cb263134b3d-kube-api-access-mbr8s" (OuterVolumeSpecName: "kube-api-access-mbr8s") pod "4d1c19bc-9031-49eb-88a1-3cb263134b3d" (UID: "4d1c19bc-9031-49eb-88a1-3cb263134b3d"). InnerVolumeSpecName "kube-api-access-mbr8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.855562 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.902938 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-utilities\") pod \"271b0494-814e-4897-9f09-d2f6b75fe66e\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903017 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-catalog-content\") pod \"271b0494-814e-4897-9f09-d2f6b75fe66e\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903063 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjs5g\" (UniqueName: \"kubernetes.io/projected/271b0494-814e-4897-9f09-d2f6b75fe66e-kube-api-access-sjs5g\") pod \"271b0494-814e-4897-9f09-d2f6b75fe66e\" (UID: \"271b0494-814e-4897-9f09-d2f6b75fe66e\") " Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903223 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903234 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903246 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903256 4675 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903266 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903276 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903287 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903295 4675 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903304 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903314 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903323 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903332 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903341 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbr8s\" (UniqueName: \"kubernetes.io/projected/4d1c19bc-9031-49eb-88a1-3cb263134b3d-kube-api-access-mbr8s\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.903350 4675 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4d1c19bc-9031-49eb-88a1-3cb263134b3d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.904068 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-utilities" (OuterVolumeSpecName: "utilities") pod "271b0494-814e-4897-9f09-d2f6b75fe66e" (UID: "271b0494-814e-4897-9f09-d2f6b75fe66e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:25 crc kubenswrapper[4675]: I0219 08:47:25.907472 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271b0494-814e-4897-9f09-d2f6b75fe66e-kube-api-access-sjs5g" (OuterVolumeSpecName: "kube-api-access-sjs5g") pod "271b0494-814e-4897-9f09-d2f6b75fe66e" (UID: "271b0494-814e-4897-9f09-d2f6b75fe66e"). InnerVolumeSpecName "kube-api-access-sjs5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.005042 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjs5g\" (UniqueName: \"kubernetes.io/projected/271b0494-814e-4897-9f09-d2f6b75fe66e-kube-api-access-sjs5g\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.005232 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.547271 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" event={"ID":"4d1c19bc-9031-49eb-88a1-3cb263134b3d","Type":"ContainerDied","Data":"c6a382c9bff1c8c41350d4e8886c3ead1f6718f0a3ac0aa737dba7cae34c8b75"} Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.547313 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-plmm7" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.547363 4675 scope.go:117] "RemoveContainer" containerID="de50c513a7f22497734c80a4bd667b0ad2672c40f7385eed2474943958cc320f" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.551565 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gr52t" event={"ID":"271b0494-814e-4897-9f09-d2f6b75fe66e","Type":"ContainerDied","Data":"17b4b03edb8fbe9ad7ebc7cada2fe02b586aae56a8505fad1b18c805f3dbe89f"} Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.551785 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gr52t" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.567944 4675 scope.go:117] "RemoveContainer" containerID="8b653f20181c4054b5443bc568c95ebf651f6a4eab883d26ae70d9fb60dfa383" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.580398 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-plmm7"] Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.583468 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-plmm7"] Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.587481 4675 scope.go:117] "RemoveContainer" containerID="ac891083f817882cf3f1f2eeabd221d187078d71799a1c2c5bc54a34f89b0249" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.602451 4675 scope.go:117] "RemoveContainer" containerID="d44eb74b19ba4bb35ca7171e9f184ba2d09e6ab4b3e77d309ac9732507e00e00" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.729145 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "271b0494-814e-4897-9f09-d2f6b75fe66e" (UID: "271b0494-814e-4897-9f09-d2f6b75fe66e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.813936 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271b0494-814e-4897-9f09-d2f6b75fe66e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.880911 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gr52t"] Feb 19 08:47:26 crc kubenswrapper[4675]: I0219 08:47:26.883373 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gr52t"] Feb 19 08:47:27 crc kubenswrapper[4675]: I0219 08:47:27.109814 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" path="/var/lib/kubelet/pods/271b0494-814e-4897-9f09-d2f6b75fe66e/volumes" Feb 19 08:47:27 crc kubenswrapper[4675]: I0219 08:47:27.110443 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d1c19bc-9031-49eb-88a1-3cb263134b3d" path="/var/lib/kubelet/pods/4d1c19bc-9031-49eb-88a1-3cb263134b3d/volumes" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.458147 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk"] Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459075 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459089 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459100 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459106 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459115 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459122 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459134 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459139 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459146 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d1c19bc-9031-49eb-88a1-3cb263134b3d" containerName="oauth-openshift" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459152 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d1c19bc-9031-49eb-88a1-3cb263134b3d" containerName="oauth-openshift" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459160 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459166 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459173 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459178 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459192 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459198 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459207 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459213 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459224 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459238 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459249 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459256 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459268 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459275 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="extract-utilities" Feb 19 08:47:30 crc kubenswrapper[4675]: E0219 08:47:30.459283 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459289 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="extract-content" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459390 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd1115a1-2835-4bc5-8d17-7387b7f34cc4" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459405 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d1c19bc-9031-49eb-88a1-3cb263134b3d" containerName="oauth-openshift" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459414 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e8bfbea-83ba-43fd-a0e6-1af01b28cbc4" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459424 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="271b0494-814e-4897-9f09-d2f6b75fe66e" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459435 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d4e815-2507-4852-b5be-40876f730e74" containerName="registry-server" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.459910 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.463037 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.463099 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.463095 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.463142 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.463286 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.463428 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.464156 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.464535 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.464736 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.481308 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.483116 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.483177 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.514802 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.515065 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.515501 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.524031 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk"] Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.560842 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561163 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f35d60c8-5897-485b-8275-670bc9bb03df-audit-dir\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561205 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc6wn\" (UniqueName: \"kubernetes.io/projected/f35d60c8-5897-485b-8275-670bc9bb03df-kube-api-access-zc6wn\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561305 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-session\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561334 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561396 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561423 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-service-ca\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561468 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561532 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561553 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561607 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-login\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561694 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-audit-policies\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561721 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-error\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.561754 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-router-certs\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662744 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662799 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f35d60c8-5897-485b-8275-670bc9bb03df-audit-dir\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662829 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc6wn\" (UniqueName: \"kubernetes.io/projected/f35d60c8-5897-485b-8275-670bc9bb03df-kube-api-access-zc6wn\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662855 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-session\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662876 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662905 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662922 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-service-ca\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662938 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662960 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662974 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.662994 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-login\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.663017 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-audit-policies\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.663033 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-error\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.663059 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-router-certs\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.664693 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-audit-policies\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.665252 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.665310 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f35d60c8-5897-485b-8275-670bc9bb03df-audit-dir\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.665908 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-service-ca\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.666059 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.669602 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-error\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.670089 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.670595 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-session\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.671144 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.672835 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.674293 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-router-certs\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.684146 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-user-template-login\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.684277 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f35d60c8-5897-485b-8275-670bc9bb03df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.690484 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc6wn\" (UniqueName: \"kubernetes.io/projected/f35d60c8-5897-485b-8275-670bc9bb03df-kube-api-access-zc6wn\") pod \"oauth-openshift-9bc7b6b6b-r6wwk\" (UID: \"f35d60c8-5897-485b-8275-670bc9bb03df\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:30 crc kubenswrapper[4675]: I0219 08:47:30.778502 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:31 crc kubenswrapper[4675]: I0219 08:47:31.017596 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk"] Feb 19 08:47:31 crc kubenswrapper[4675]: W0219 08:47:31.029475 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf35d60c8_5897_485b_8275_670bc9bb03df.slice/crio-20c6a580e241fd97098a41012e18ce9f67efefbf806e2d649217370776440d6d WatchSource:0}: Error finding container 20c6a580e241fd97098a41012e18ce9f67efefbf806e2d649217370776440d6d: Status 404 returned error can't find the container with id 20c6a580e241fd97098a41012e18ce9f67efefbf806e2d649217370776440d6d Feb 19 08:47:31 crc kubenswrapper[4675]: I0219 08:47:31.580320 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" event={"ID":"f35d60c8-5897-485b-8275-670bc9bb03df","Type":"ContainerStarted","Data":"f9b9cc64e8a9898e4954ad1b9fcfe7b0a5b4fd37838ec7a9726958c9e0f879fb"} Feb 19 08:47:31 crc kubenswrapper[4675]: I0219 08:47:31.580672 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" event={"ID":"f35d60c8-5897-485b-8275-670bc9bb03df","Type":"ContainerStarted","Data":"20c6a580e241fd97098a41012e18ce9f67efefbf806e2d649217370776440d6d"} Feb 19 08:47:31 crc kubenswrapper[4675]: I0219 08:47:31.580762 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:31 crc kubenswrapper[4675]: I0219 08:47:31.601448 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" podStartSLOduration=32.6014266 podStartE2EDuration="32.6014266s" podCreationTimestamp="2026-02-19 08:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:47:31.597287626 +0000 UTC m=+233.224377894" watchObservedRunningTime="2026-02-19 08:47:31.6014266 +0000 UTC m=+233.228516868" Feb 19 08:47:31 crc kubenswrapper[4675]: I0219 08:47:31.720370 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-r6wwk" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.004056 4675 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.004784 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.005924 4675 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.006193 4675 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.006584 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27" gracePeriod=15 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.006604 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced" gracePeriod=15 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.006753 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0" gracePeriod=15 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.006754 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870" gracePeriod=15 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.006813 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830" gracePeriod=15 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009045 4675 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.009232 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009247 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.009259 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009267 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.009275 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009282 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.009292 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009300 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.009313 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009320 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.009332 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009339 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.009348 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009355 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009466 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009479 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009486 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009495 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009507 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.009756 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.053323 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179140 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179211 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179235 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179488 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179542 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179615 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179681 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.179700 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281194 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281283 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281338 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281395 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281444 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281514 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281545 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.281604 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.282410 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.282543 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.282902 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.282915 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.283293 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.283474 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.283567 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.283612 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.334538 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:47:32 crc kubenswrapper[4675]: W0219 08:47:32.364690 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-2137abaf8aac7c3ec43fa7c43f18818b1f484df2666fc82d2cddbb905c44f84b WatchSource:0}: Error finding container 2137abaf8aac7c3ec43fa7c43f18818b1f484df2666fc82d2cddbb905c44f84b: Status 404 returned error can't find the container with id 2137abaf8aac7c3ec43fa7c43f18818b1f484df2666fc82d2cddbb905c44f84b Feb 19 08:47:32 crc kubenswrapper[4675]: E0219 08:47:32.367453 4675 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895998b622ecfbe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-19 08:47:32.366897086 +0000 UTC m=+233.993987354,LastTimestamp:2026-02-19 08:47:32.366897086 +0000 UTC m=+233.993987354,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.588038 4675 generic.go:334] "Generic (PLEG): container finished" podID="78e84b11-6e92-4596-9397-017767524db7" containerID="e7a94ddb33287eccfc95bb313522f7d43d5206d8127b7d83b8f9a70154c44ed7" exitCode=0 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.588118 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"78e84b11-6e92-4596-9397-017767524db7","Type":"ContainerDied","Data":"e7a94ddb33287eccfc95bb313522f7d43d5206d8127b7d83b8f9a70154c44ed7"} Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.588799 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.589040 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.589221 4675 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.590663 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2137abaf8aac7c3ec43fa7c43f18818b1f484df2666fc82d2cddbb905c44f84b"} Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.592423 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.593391 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.594019 4675 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced" exitCode=0 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.594041 4675 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830" exitCode=0 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.594048 4675 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870" exitCode=0 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.594055 4675 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0" exitCode=2 Feb 19 08:47:32 crc kubenswrapper[4675]: I0219 08:47:32.594429 4675 scope.go:117] "RemoveContainer" containerID="456975b8ece15bb0fe1ac4dbe02ef908f8d3cf4a07d91106adc2ee891404eeb0" Feb 19 08:47:33 crc kubenswrapper[4675]: I0219 08:47:33.604275 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00"} Feb 19 08:47:33 crc kubenswrapper[4675]: I0219 08:47:33.605319 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:33 crc kubenswrapper[4675]: I0219 08:47:33.606180 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:33 crc kubenswrapper[4675]: I0219 08:47:33.607242 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 19 08:47:33 crc kubenswrapper[4675]: I0219 08:47:33.830823 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:47:33 crc kubenswrapper[4675]: I0219 08:47:33.831346 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:33 crc kubenswrapper[4675]: I0219 08:47:33.831540 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.004176 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78e84b11-6e92-4596-9397-017767524db7-kube-api-access\") pod \"78e84b11-6e92-4596-9397-017767524db7\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.004241 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-var-lock\") pod \"78e84b11-6e92-4596-9397-017767524db7\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.004289 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-kubelet-dir\") pod \"78e84b11-6e92-4596-9397-017767524db7\" (UID: \"78e84b11-6e92-4596-9397-017767524db7\") " Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.004372 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-var-lock" (OuterVolumeSpecName: "var-lock") pod "78e84b11-6e92-4596-9397-017767524db7" (UID: "78e84b11-6e92-4596-9397-017767524db7"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.004493 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "78e84b11-6e92-4596-9397-017767524db7" (UID: "78e84b11-6e92-4596-9397-017767524db7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.004714 4675 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-var-lock\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.004730 4675 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78e84b11-6e92-4596-9397-017767524db7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.010057 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78e84b11-6e92-4596-9397-017767524db7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "78e84b11-6e92-4596-9397-017767524db7" (UID: "78e84b11-6e92-4596-9397-017767524db7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.105571 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78e84b11-6e92-4596-9397-017767524db7-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:34 crc kubenswrapper[4675]: E0219 08:47:34.206958 4675 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" volumeName="registry-storage" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.432150 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.433605 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.434333 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.434877 4675 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.435137 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.611359 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.611463 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.611542 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.611572 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.611596 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.611669 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.612165 4675 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.612181 4675 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.612191 4675 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.619303 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.620164 4675 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27" exitCode=0 Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.620261 4675 scope.go:117] "RemoveContainer" containerID="e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.620263 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.622386 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.624792 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"78e84b11-6e92-4596-9397-017767524db7","Type":"ContainerDied","Data":"65eb77593ef4d467e6644496d5fa6015089e4e45a65cac7a3a8b4b363dbc87eb"} Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.624857 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65eb77593ef4d467e6644496d5fa6015089e4e45a65cac7a3a8b4b363dbc87eb" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.635857 4675 scope.go:117] "RemoveContainer" containerID="9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.637463 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.638090 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.638323 4675 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.638749 4675 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.639042 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.639269 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.646751 4675 scope.go:117] "RemoveContainer" containerID="90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.660398 4675 scope.go:117] "RemoveContainer" containerID="b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.674322 4675 scope.go:117] "RemoveContainer" containerID="ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.685801 4675 scope.go:117] "RemoveContainer" containerID="85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.701735 4675 scope.go:117] "RemoveContainer" containerID="e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced" Feb 19 08:47:34 crc kubenswrapper[4675]: E0219 08:47:34.702126 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\": container with ID starting with e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced not found: ID does not exist" containerID="e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.702170 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced"} err="failed to get container status \"e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\": rpc error: code = NotFound desc = could not find container \"e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced\": container with ID starting with e8e9f7db76a867ec28916d0499a206e88d181f0018a7561fac0b046b4ce5fced not found: ID does not exist" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.702201 4675 scope.go:117] "RemoveContainer" containerID="9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830" Feb 19 08:47:34 crc kubenswrapper[4675]: E0219 08:47:34.702788 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\": container with ID starting with 9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830 not found: ID does not exist" containerID="9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.702818 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830"} err="failed to get container status \"9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\": rpc error: code = NotFound desc = could not find container \"9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830\": container with ID starting with 9453ae113c1831a2b3b0bc758a87ed021e35e8511818fddf596fc6fee8e07830 not found: ID does not exist" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.702844 4675 scope.go:117] "RemoveContainer" containerID="90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870" Feb 19 08:47:34 crc kubenswrapper[4675]: E0219 08:47:34.703107 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\": container with ID starting with 90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870 not found: ID does not exist" containerID="90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.703133 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870"} err="failed to get container status \"90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\": rpc error: code = NotFound desc = could not find container \"90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870\": container with ID starting with 90c8619438862ee51f6e582ea8ac772f1a3b3302dfb691461e7deaff368c5870 not found: ID does not exist" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.703151 4675 scope.go:117] "RemoveContainer" containerID="b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0" Feb 19 08:47:34 crc kubenswrapper[4675]: E0219 08:47:34.703450 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\": container with ID starting with b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0 not found: ID does not exist" containerID="b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.703480 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0"} err="failed to get container status \"b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\": rpc error: code = NotFound desc = could not find container \"b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0\": container with ID starting with b8a2365d8f8ea82552df9321215f23e8b3b381e7430d39d9dad84391962a33e0 not found: ID does not exist" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.703499 4675 scope.go:117] "RemoveContainer" containerID="ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27" Feb 19 08:47:34 crc kubenswrapper[4675]: E0219 08:47:34.704327 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\": container with ID starting with ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27 not found: ID does not exist" containerID="ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.704368 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27"} err="failed to get container status \"ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\": rpc error: code = NotFound desc = could not find container \"ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27\": container with ID starting with ab3b30f522dc3f24060828e18d8cc80fa134f5c2d1aa0f8fae369ac002792a27 not found: ID does not exist" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.704387 4675 scope.go:117] "RemoveContainer" containerID="85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a" Feb 19 08:47:34 crc kubenswrapper[4675]: E0219 08:47:34.704715 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\": container with ID starting with 85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a not found: ID does not exist" containerID="85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a" Feb 19 08:47:34 crc kubenswrapper[4675]: I0219 08:47:34.704943 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a"} err="failed to get container status \"85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\": rpc error: code = NotFound desc = could not find container \"85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a\": container with ID starting with 85a48489b85c1eea538e2c8c69d4156910419bdfc1f07f9709e95d785317218a not found: ID does not exist" Feb 19 08:47:35 crc kubenswrapper[4675]: I0219 08:47:35.112574 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 19 08:47:35 crc kubenswrapper[4675]: E0219 08:47:35.195449 4675 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895998b622ecfbe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-19 08:47:32.366897086 +0000 UTC m=+233.993987354,LastTimestamp:2026-02-19 08:47:32.366897086 +0000 UTC m=+233.993987354,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.370591 4675 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.371239 4675 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.371480 4675 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.371866 4675 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.373452 4675 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:38 crc kubenswrapper[4675]: I0219 08:47:38.373495 4675 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.373863 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="200ms" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.575521 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="400ms" Feb 19 08:47:38 crc kubenswrapper[4675]: E0219 08:47:38.976380 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="800ms" Feb 19 08:47:39 crc kubenswrapper[4675]: I0219 08:47:39.109055 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:39 crc kubenswrapper[4675]: I0219 08:47:39.109721 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:39 crc kubenswrapper[4675]: E0219 08:47:39.778035 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="1.6s" Feb 19 08:47:40 crc kubenswrapper[4675]: E0219 08:47:40.964248 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:47:40Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:47:40Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:47:40Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-19T08:47:40Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:40 crc kubenswrapper[4675]: E0219 08:47:40.965324 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:40 crc kubenswrapper[4675]: E0219 08:47:40.965962 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:40 crc kubenswrapper[4675]: E0219 08:47:40.966422 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:40 crc kubenswrapper[4675]: E0219 08:47:40.966703 4675 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:40 crc kubenswrapper[4675]: E0219 08:47:40.966740 4675 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 19 08:47:41 crc kubenswrapper[4675]: E0219 08:47:41.380872 4675 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="3.2s" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.103509 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.104560 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.105327 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.125006 4675 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.125066 4675 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:43 crc kubenswrapper[4675]: E0219 08:47:43.125574 4675 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.126184 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:43 crc kubenswrapper[4675]: W0219 08:47:43.152311 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-cb6f3e2aa5fed1d67989c9f3aba451340b7f6541285a5558a9adf07a20487933 WatchSource:0}: Error finding container cb6f3e2aa5fed1d67989c9f3aba451340b7f6541285a5558a9adf07a20487933: Status 404 returned error can't find the container with id cb6f3e2aa5fed1d67989c9f3aba451340b7f6541285a5558a9adf07a20487933 Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.671957 4675 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f11717f6cdbf431c16580ca0bb735a2d8bcd7a7ddda8b6985821825f7cd5769d" exitCode=0 Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.672031 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f11717f6cdbf431c16580ca0bb735a2d8bcd7a7ddda8b6985821825f7cd5769d"} Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.672092 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cb6f3e2aa5fed1d67989c9f3aba451340b7f6541285a5558a9adf07a20487933"} Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.672683 4675 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.672733 4675 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.673402 4675 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:43 crc kubenswrapper[4675]: E0219 08:47:43.673476 4675 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:43 crc kubenswrapper[4675]: I0219 08:47:43.673917 4675 status_manager.go:851] "Failed to get status for pod" podUID="78e84b11-6e92-4596-9397-017767524db7" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 19 08:47:44 crc kubenswrapper[4675]: I0219 08:47:44.683212 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cc453eb5eee56f7369c462f0e94f9b07a42d0172f36ced3adbab7e39ecd9fb3d"} Feb 19 08:47:44 crc kubenswrapper[4675]: I0219 08:47:44.683602 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5ba24493b896b3ea0b6eda6e695e56810a8d811e9cb06e039714ddc11a9ee055"} Feb 19 08:47:44 crc kubenswrapper[4675]: I0219 08:47:44.683613 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2c7dc967ef31ac8870887237fce92b6f55687c9a809cef2b2e3581be3a652fc3"} Feb 19 08:47:44 crc kubenswrapper[4675]: I0219 08:47:44.683623 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7d0cfe1bdb26982455dd89eb4131589a99c62e3ee95b95fe3f6f47c21dc4fed9"} Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.692626 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.692708 4675 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d" exitCode=1 Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.692793 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d"} Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.693452 4675 scope.go:117] "RemoveContainer" containerID="a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d" Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.697307 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"30edb62920fc1fbdab4fcbc976d0ac40de0d21bc7f72921373871ba2298814b8"} Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.697692 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.697814 4675 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:45 crc kubenswrapper[4675]: I0219 08:47:45.697840 4675 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:46 crc kubenswrapper[4675]: I0219 08:47:46.706352 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 19 08:47:46 crc kubenswrapper[4675]: I0219 08:47:46.706707 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6a168ea972232e452f018e740876984d8a30b89908bdd1300f50aa4607a98ec7"} Feb 19 08:47:46 crc kubenswrapper[4675]: I0219 08:47:46.726076 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:47:48 crc kubenswrapper[4675]: I0219 08:47:48.126933 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:48 crc kubenswrapper[4675]: I0219 08:47:48.127440 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:48 crc kubenswrapper[4675]: I0219 08:47:48.172155 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:48 crc kubenswrapper[4675]: I0219 08:47:48.233400 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:47:48 crc kubenswrapper[4675]: I0219 08:47:48.233716 4675 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 19 08:47:48 crc kubenswrapper[4675]: I0219 08:47:48.233789 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 19 08:47:50 crc kubenswrapper[4675]: I0219 08:47:50.709056 4675 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:50 crc kubenswrapper[4675]: I0219 08:47:50.727207 4675 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:50 crc kubenswrapper[4675]: I0219 08:47:50.727241 4675 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:50 crc kubenswrapper[4675]: I0219 08:47:50.730844 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:47:50 crc kubenswrapper[4675]: I0219 08:47:50.806055 4675 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="598bd1ef-5ef2-4050-a0cb-0d7bb45d73fa" Feb 19 08:47:51 crc kubenswrapper[4675]: I0219 08:47:51.734584 4675 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:51 crc kubenswrapper[4675]: I0219 08:47:51.734674 4675 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ffd5e4b9-c1df-47cf-bf03-8596cdbd375e" Feb 19 08:47:51 crc kubenswrapper[4675]: I0219 08:47:51.739434 4675 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="598bd1ef-5ef2-4050-a0cb-0d7bb45d73fa" Feb 19 08:47:56 crc kubenswrapper[4675]: I0219 08:47:56.728968 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 19 08:47:56 crc kubenswrapper[4675]: I0219 08:47:56.982814 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 19 08:47:57 crc kubenswrapper[4675]: I0219 08:47:57.424307 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 19 08:47:57 crc kubenswrapper[4675]: I0219 08:47:57.559307 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 19 08:47:57 crc kubenswrapper[4675]: I0219 08:47:57.948110 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 19 08:47:58 crc kubenswrapper[4675]: I0219 08:47:58.149683 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 19 08:47:58 crc kubenswrapper[4675]: I0219 08:47:58.177706 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 19 08:47:58 crc kubenswrapper[4675]: I0219 08:47:58.234204 4675 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 19 08:47:58 crc kubenswrapper[4675]: I0219 08:47:58.234321 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 19 08:47:58 crc kubenswrapper[4675]: I0219 08:47:58.809901 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 19 08:47:59 crc kubenswrapper[4675]: I0219 08:47:59.165298 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 19 08:47:59 crc kubenswrapper[4675]: I0219 08:47:59.280892 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 19 08:47:59 crc kubenswrapper[4675]: I0219 08:47:59.359515 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 19 08:47:59 crc kubenswrapper[4675]: I0219 08:47:59.562995 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 19 08:47:59 crc kubenswrapper[4675]: I0219 08:47:59.869738 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 19 08:48:00 crc kubenswrapper[4675]: I0219 08:48:00.363218 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 19 08:48:01 crc kubenswrapper[4675]: I0219 08:48:01.038497 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 19 08:48:01 crc kubenswrapper[4675]: I0219 08:48:01.307545 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 19 08:48:02 crc kubenswrapper[4675]: I0219 08:48:02.196577 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 19 08:48:02 crc kubenswrapper[4675]: I0219 08:48:02.475234 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 19 08:48:02 crc kubenswrapper[4675]: I0219 08:48:02.825076 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.015155 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.065299 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.125198 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.164804 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.341268 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.377985 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.554900 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 19 08:48:03 crc kubenswrapper[4675]: I0219 08:48:03.723874 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 19 08:48:04 crc kubenswrapper[4675]: I0219 08:48:04.283581 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 19 08:48:04 crc kubenswrapper[4675]: I0219 08:48:04.381695 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 19 08:48:04 crc kubenswrapper[4675]: I0219 08:48:04.627796 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 19 08:48:04 crc kubenswrapper[4675]: I0219 08:48:04.950877 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 19 08:48:04 crc kubenswrapper[4675]: I0219 08:48:04.996166 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.028597 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.075018 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.179215 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.517842 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.523594 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.556610 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.579040 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.744747 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.793405 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.920119 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.967097 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 19 08:48:05 crc kubenswrapper[4675]: I0219 08:48:05.981531 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 19 08:48:06 crc kubenswrapper[4675]: I0219 08:48:06.334965 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 19 08:48:06 crc kubenswrapper[4675]: I0219 08:48:06.420220 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 19 08:48:06 crc kubenswrapper[4675]: I0219 08:48:06.421324 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 19 08:48:06 crc kubenswrapper[4675]: I0219 08:48:06.757887 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 19 08:48:06 crc kubenswrapper[4675]: I0219 08:48:06.857573 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 19 08:48:06 crc kubenswrapper[4675]: I0219 08:48:06.928825 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.019140 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.290667 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.356906 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.402891 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.561228 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.564784 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.608967 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.649254 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.715650 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.746941 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 19 08:48:07 crc kubenswrapper[4675]: I0219 08:48:07.822191 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.062718 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.211223 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.216727 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.230676 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.233895 4675 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.233962 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.234024 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.234764 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"6a168ea972232e452f018e740876984d8a30b89908bdd1300f50aa4607a98ec7"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.234921 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://6a168ea972232e452f018e740876984d8a30b89908bdd1300f50aa4607a98ec7" gracePeriod=30 Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.239097 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.241813 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.272008 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.318412 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.347849 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.416265 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.452046 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.464898 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.491006 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.571793 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.656494 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.812421 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.839063 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 19 08:48:08 crc kubenswrapper[4675]: I0219 08:48:08.899210 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.041981 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.047689 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.092161 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.168677 4675 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.227143 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.343697 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.370511 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.371380 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.373972 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.403772 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.415948 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.497546 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.524200 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.524732 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.560512 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.566302 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.571627 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.573242 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.695327 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.697009 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.738059 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.827098 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.827158 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.876360 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 19 08:48:09 crc kubenswrapper[4675]: I0219 08:48:09.977457 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.121991 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.125742 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.255961 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.284825 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.314236 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.318211 4675 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.319121 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=38.319106104 podStartE2EDuration="38.319106104s" podCreationTimestamp="2026-02-19 08:47:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:47:50.752435269 +0000 UTC m=+252.379525547" watchObservedRunningTime="2026-02-19 08:48:10.319106104 +0000 UTC m=+271.946196372" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.322124 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.322168 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.326682 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.338685 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.338666015 podStartE2EDuration="20.338666015s" podCreationTimestamp="2026-02-19 08:47:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:48:10.338315157 +0000 UTC m=+271.965405425" watchObservedRunningTime="2026-02-19 08:48:10.338666015 +0000 UTC m=+271.965756283" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.377094 4675 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.533442 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.540519 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.562158 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.629713 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.656203 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.721758 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.767964 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.778242 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.889397 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.950615 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.958323 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 19 08:48:10 crc kubenswrapper[4675]: I0219 08:48:10.967101 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.074960 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.150112 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.297175 4675 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.332058 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.361701 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.372908 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.394835 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.561734 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.582566 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.645923 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.692373 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.724005 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.741661 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.742549 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.749035 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.822257 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.901079 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.919094 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 19 08:48:11 crc kubenswrapper[4675]: I0219 08:48:11.942525 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.060407 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.077750 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.092113 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.094507 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.167334 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.170688 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.358950 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.379932 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.389596 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.408598 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.490949 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.525876 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.576020 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.662681 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.679875 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.757893 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.758984 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.771180 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.783577 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.978840 4675 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 19 08:48:12 crc kubenswrapper[4675]: I0219 08:48:12.979089 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00" gracePeriod=5 Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.149460 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.190549 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.327202 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.399992 4675 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.413136 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.422719 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.508754 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.552043 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.557957 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.573853 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.585853 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.686233 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.712705 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.715892 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.742228 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.791191 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.845568 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.949428 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 19 08:48:13 crc kubenswrapper[4675]: I0219 08:48:13.980744 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.049863 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.093604 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.189657 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.240225 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.264354 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.338471 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.345548 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.406164 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.539917 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.626578 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.720684 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.747794 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.837450 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.874173 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.967746 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 19 08:48:14 crc kubenswrapper[4675]: I0219 08:48:14.977285 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.053448 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.078423 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.148921 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.153456 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.204389 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.210240 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.231869 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.320974 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.361712 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.369369 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.403980 4675 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.451301 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.456885 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.501998 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.584631 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.595679 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.726290 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.765765 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.818823 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.836969 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.850914 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 19 08:48:15 crc kubenswrapper[4675]: I0219 08:48:15.904059 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.019545 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.198479 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.335837 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.440430 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.512223 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.520053 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.648393 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.825098 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.830347 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 19 08:48:16 crc kubenswrapper[4675]: I0219 08:48:16.863168 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.070273 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.096692 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.260335 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.394941 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.476370 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.574581 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.651403 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.756678 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 19 08:48:17 crc kubenswrapper[4675]: I0219 08:48:17.817330 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.020769 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.061360 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.351589 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.452631 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.560014 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.560140 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.566427 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.695748 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.695806 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696044 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696048 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696155 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696197 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696037 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696322 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696373 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696749 4675 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696783 4675 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696800 4675 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.696817 4675 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.703860 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.797326 4675 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 19 08:48:18 crc kubenswrapper[4675]: I0219 08:48:18.985040 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.012147 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.012212 4675 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00" exitCode=137 Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.012267 4675 scope.go:117] "RemoveContainer" containerID="a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.012399 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.035391 4675 scope.go:117] "RemoveContainer" containerID="a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00" Feb 19 08:48:19 crc kubenswrapper[4675]: E0219 08:48:19.036125 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00\": container with ID starting with a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00 not found: ID does not exist" containerID="a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.036252 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00"} err="failed to get container status \"a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00\": rpc error: code = NotFound desc = could not find container \"a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00\": container with ID starting with a92efad89463f884c10010ddf6315810b88b7b77e82dca349ce612dc24effc00 not found: ID does not exist" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.110683 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.111259 4675 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.122574 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.122625 4675 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="4c1e250a-1ca9-4102-bd9a-4920308f8abd" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.126245 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.126350 4675 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="4c1e250a-1ca9-4102-bd9a-4920308f8abd" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.207142 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.354326 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.358318 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.397159 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 19 08:48:19 crc kubenswrapper[4675]: I0219 08:48:19.821627 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 19 08:48:20 crc kubenswrapper[4675]: I0219 08:48:20.379951 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 19 08:48:20 crc kubenswrapper[4675]: I0219 08:48:20.486434 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 19 08:48:20 crc kubenswrapper[4675]: I0219 08:48:20.746769 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 19 08:48:20 crc kubenswrapper[4675]: I0219 08:48:20.755455 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 19 08:48:38 crc kubenswrapper[4675]: I0219 08:48:38.868451 4675 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 19 08:48:39 crc kubenswrapper[4675]: I0219 08:48:39.113383 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Feb 19 08:48:39 crc kubenswrapper[4675]: I0219 08:48:39.115368 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 19 08:48:39 crc kubenswrapper[4675]: I0219 08:48:39.115438 4675 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6a168ea972232e452f018e740876984d8a30b89908bdd1300f50aa4607a98ec7" exitCode=137 Feb 19 08:48:39 crc kubenswrapper[4675]: I0219 08:48:39.115481 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6a168ea972232e452f018e740876984d8a30b89908bdd1300f50aa4607a98ec7"} Feb 19 08:48:39 crc kubenswrapper[4675]: I0219 08:48:39.115520 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3914f25a0174d078cad81302ef97e9133881dd4bcc89d5afbd22d4af31aeda5e"} Feb 19 08:48:39 crc kubenswrapper[4675]: I0219 08:48:39.115544 4675 scope.go:117] "RemoveContainer" containerID="a766ec85d32dce92585bc2a5983111cdb3f45ba858595e91cd570f7a815b4d3d" Feb 19 08:48:40 crc kubenswrapper[4675]: I0219 08:48:40.122504 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Feb 19 08:48:40 crc kubenswrapper[4675]: I0219 08:48:40.125359 4675 generic.go:334] "Generic (PLEG): container finished" podID="017fb59c-7331-42bf-9a21-8dff74814740" containerID="890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8" exitCode=0 Feb 19 08:48:40 crc kubenswrapper[4675]: I0219 08:48:40.125584 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" event={"ID":"017fb59c-7331-42bf-9a21-8dff74814740","Type":"ContainerDied","Data":"890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8"} Feb 19 08:48:40 crc kubenswrapper[4675]: I0219 08:48:40.125870 4675 scope.go:117] "RemoveContainer" containerID="890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8" Feb 19 08:48:41 crc kubenswrapper[4675]: I0219 08:48:41.134391 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" event={"ID":"017fb59c-7331-42bf-9a21-8dff74814740","Type":"ContainerStarted","Data":"e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862"} Feb 19 08:48:41 crc kubenswrapper[4675]: I0219 08:48:41.135183 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:48:41 crc kubenswrapper[4675]: I0219 08:48:41.140678 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:48:46 crc kubenswrapper[4675]: I0219 08:48:46.726608 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:48:48 crc kubenswrapper[4675]: I0219 08:48:48.234023 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:48:48 crc kubenswrapper[4675]: I0219 08:48:48.240513 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:48:49 crc kubenswrapper[4675]: I0219 08:48:49.177643 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 19 08:48:59 crc kubenswrapper[4675]: I0219 08:48:59.503367 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-srrlc"] Feb 19 08:48:59 crc kubenswrapper[4675]: I0219 08:48:59.504409 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" podUID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" containerName="controller-manager" containerID="cri-o://51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92" gracePeriod=30 Feb 19 08:48:59 crc kubenswrapper[4675]: I0219 08:48:59.506314 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q"] Feb 19 08:48:59 crc kubenswrapper[4675]: I0219 08:48:59.506456 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" podUID="366baa67-abf1-485c-9430-1a891c9e58c7" containerName="route-controller-manager" containerID="cri-o://b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c" gracePeriod=30 Feb 19 08:48:59 crc kubenswrapper[4675]: I0219 08:48:59.937543 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:48:59 crc kubenswrapper[4675]: I0219 08:48:59.944926 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009486 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4796\" (UniqueName: \"kubernetes.io/projected/366baa67-abf1-485c-9430-1a891c9e58c7-kube-api-access-s4796\") pod \"366baa67-abf1-485c-9430-1a891c9e58c7\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009549 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/366baa67-abf1-485c-9430-1a891c9e58c7-serving-cert\") pod \"366baa67-abf1-485c-9430-1a891c9e58c7\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009592 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-serving-cert\") pod \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009630 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-config\") pod \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009683 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-config\") pod \"366baa67-abf1-485c-9430-1a891c9e58c7\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009707 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-client-ca\") pod \"366baa67-abf1-485c-9430-1a891c9e58c7\" (UID: \"366baa67-abf1-485c-9430-1a891c9e58c7\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009737 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-client-ca\") pod \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009765 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbjvq\" (UniqueName: \"kubernetes.io/projected/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-kube-api-access-vbjvq\") pod \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.009790 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-proxy-ca-bundles\") pod \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\" (UID: \"ad9383f1-0293-4f42-b7f9-2e3fef0e4569\") " Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.010508 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ad9383f1-0293-4f42-b7f9-2e3fef0e4569" (UID: "ad9383f1-0293-4f42-b7f9-2e3fef0e4569"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.010517 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-client-ca" (OuterVolumeSpecName: "client-ca") pod "ad9383f1-0293-4f42-b7f9-2e3fef0e4569" (UID: "ad9383f1-0293-4f42-b7f9-2e3fef0e4569"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.010776 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-client-ca" (OuterVolumeSpecName: "client-ca") pod "366baa67-abf1-485c-9430-1a891c9e58c7" (UID: "366baa67-abf1-485c-9430-1a891c9e58c7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.010811 4675 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.010891 4675 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-client-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.010794 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-config" (OuterVolumeSpecName: "config") pod "366baa67-abf1-485c-9430-1a891c9e58c7" (UID: "366baa67-abf1-485c-9430-1a891c9e58c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.011628 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-config" (OuterVolumeSpecName: "config") pod "ad9383f1-0293-4f42-b7f9-2e3fef0e4569" (UID: "ad9383f1-0293-4f42-b7f9-2e3fef0e4569"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.017318 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366baa67-abf1-485c-9430-1a891c9e58c7-kube-api-access-s4796" (OuterVolumeSpecName: "kube-api-access-s4796") pod "366baa67-abf1-485c-9430-1a891c9e58c7" (UID: "366baa67-abf1-485c-9430-1a891c9e58c7"). InnerVolumeSpecName "kube-api-access-s4796". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.017331 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366baa67-abf1-485c-9430-1a891c9e58c7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "366baa67-abf1-485c-9430-1a891c9e58c7" (UID: "366baa67-abf1-485c-9430-1a891c9e58c7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.018064 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-kube-api-access-vbjvq" (OuterVolumeSpecName: "kube-api-access-vbjvq") pod "ad9383f1-0293-4f42-b7f9-2e3fef0e4569" (UID: "ad9383f1-0293-4f42-b7f9-2e3fef0e4569"). InnerVolumeSpecName "kube-api-access-vbjvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.018250 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ad9383f1-0293-4f42-b7f9-2e3fef0e4569" (UID: "ad9383f1-0293-4f42-b7f9-2e3fef0e4569"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.111968 4675 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-client-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.112027 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbjvq\" (UniqueName: \"kubernetes.io/projected/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-kube-api-access-vbjvq\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.112039 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4796\" (UniqueName: \"kubernetes.io/projected/366baa67-abf1-485c-9430-1a891c9e58c7-kube-api-access-s4796\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.112048 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/366baa67-abf1-485c-9430-1a891c9e58c7-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.112060 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.112075 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9383f1-0293-4f42-b7f9-2e3fef0e4569-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.112089 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366baa67-abf1-485c-9430-1a891c9e58c7-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.229799 4675 generic.go:334] "Generic (PLEG): container finished" podID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" containerID="51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92" exitCode=0 Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.229908 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" event={"ID":"ad9383f1-0293-4f42-b7f9-2e3fef0e4569","Type":"ContainerDied","Data":"51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92"} Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.229927 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.230060 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-srrlc" event={"ID":"ad9383f1-0293-4f42-b7f9-2e3fef0e4569","Type":"ContainerDied","Data":"3922889b5e9e5ebe6392c49323b691eec8a8ede0ae2457bfd3d48c164ab159d5"} Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.230081 4675 scope.go:117] "RemoveContainer" containerID="51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.233347 4675 generic.go:334] "Generic (PLEG): container finished" podID="366baa67-abf1-485c-9430-1a891c9e58c7" containerID="b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c" exitCode=0 Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.233381 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.233387 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" event={"ID":"366baa67-abf1-485c-9430-1a891c9e58c7","Type":"ContainerDied","Data":"b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c"} Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.233550 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q" event={"ID":"366baa67-abf1-485c-9430-1a891c9e58c7","Type":"ContainerDied","Data":"54113fdc183c119482cb7c7bc4855a3bdecaee5ccaf07fbc9d78ffcc5feaee3a"} Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.252905 4675 scope.go:117] "RemoveContainer" containerID="51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92" Feb 19 08:49:00 crc kubenswrapper[4675]: E0219 08:49:00.253392 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92\": container with ID starting with 51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92 not found: ID does not exist" containerID="51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.253430 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92"} err="failed to get container status \"51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92\": rpc error: code = NotFound desc = could not find container \"51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92\": container with ID starting with 51477a660db6a3d48af9a1e1361820ce26fb5683313efe715d90c1886f77eb92 not found: ID does not exist" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.253453 4675 scope.go:117] "RemoveContainer" containerID="b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.257736 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-srrlc"] Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.263102 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-srrlc"] Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.270005 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q"] Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.274016 4675 scope.go:117] "RemoveContainer" containerID="b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.274238 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rrj8q"] Feb 19 08:49:00 crc kubenswrapper[4675]: E0219 08:49:00.274452 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c\": container with ID starting with b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c not found: ID does not exist" containerID="b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c" Feb 19 08:49:00 crc kubenswrapper[4675]: I0219 08:49:00.274486 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c"} err="failed to get container status \"b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c\": rpc error: code = NotFound desc = could not find container \"b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c\": container with ID starting with b9a59f0f3d1dfd38cefae767aeaf57a90de4ea65ec1532a16a488f7f6f15eb3c not found: ID does not exist" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.109859 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366baa67-abf1-485c-9430-1a891c9e58c7" path="/var/lib/kubelet/pods/366baa67-abf1-485c-9430-1a891c9e58c7/volumes" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.110700 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" path="/var/lib/kubelet/pods/ad9383f1-0293-4f42-b7f9-2e3fef0e4569/volumes" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509188 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-54f57c75f8-cnbng"] Feb 19 08:49:01 crc kubenswrapper[4675]: E0219 08:49:01.509520 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366baa67-abf1-485c-9430-1a891c9e58c7" containerName="route-controller-manager" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509539 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="366baa67-abf1-485c-9430-1a891c9e58c7" containerName="route-controller-manager" Feb 19 08:49:01 crc kubenswrapper[4675]: E0219 08:49:01.509565 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78e84b11-6e92-4596-9397-017767524db7" containerName="installer" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509573 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="78e84b11-6e92-4596-9397-017767524db7" containerName="installer" Feb 19 08:49:01 crc kubenswrapper[4675]: E0219 08:49:01.509581 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" containerName="controller-manager" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509588 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" containerName="controller-manager" Feb 19 08:49:01 crc kubenswrapper[4675]: E0219 08:49:01.509596 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509609 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509735 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9383f1-0293-4f42-b7f9-2e3fef0e4569" containerName="controller-manager" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509745 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="78e84b11-6e92-4596-9397-017767524db7" containerName="installer" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509758 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="366baa67-abf1-485c-9430-1a891c9e58c7" containerName="route-controller-manager" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.509768 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.510193 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.512068 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.512242 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.512535 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn"] Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.513332 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.515130 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.515695 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.515740 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.515496 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.516117 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.516557 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.516581 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.516569 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.516655 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.517662 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.526128 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54f57c75f8-cnbng"] Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.526836 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.529411 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn"] Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.529510 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmxsb\" (UniqueName: \"kubernetes.io/projected/f95c097e-87aa-4cf6-b1f9-a606463c406e-kube-api-access-wmxsb\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.529781 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f95c097e-87aa-4cf6-b1f9-a606463c406e-serving-cert\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.529923 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-config\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.530066 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-proxy-ca-bundles\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.530210 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-client-ca\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.530283 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb9wm\" (UniqueName: \"kubernetes.io/projected/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-kube-api-access-mb9wm\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.530333 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-serving-cert\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.530381 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-config\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.530401 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-client-ca\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.631736 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-proxy-ca-bundles\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.631790 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-client-ca\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.631816 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb9wm\" (UniqueName: \"kubernetes.io/projected/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-kube-api-access-mb9wm\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.632044 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-serving-cert\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.632141 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-config\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.632169 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-client-ca\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.632401 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmxsb\" (UniqueName: \"kubernetes.io/projected/f95c097e-87aa-4cf6-b1f9-a606463c406e-kube-api-access-wmxsb\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.632504 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f95c097e-87aa-4cf6-b1f9-a606463c406e-serving-cert\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.632553 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-config\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.633225 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-client-ca\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.633429 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-proxy-ca-bundles\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.634036 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-client-ca\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.634067 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-config\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.635223 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f95c097e-87aa-4cf6-b1f9-a606463c406e-config\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.647494 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f95c097e-87aa-4cf6-b1f9-a606463c406e-serving-cert\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.651627 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmxsb\" (UniqueName: \"kubernetes.io/projected/f95c097e-87aa-4cf6-b1f9-a606463c406e-kube-api-access-wmxsb\") pod \"controller-manager-54f57c75f8-cnbng\" (UID: \"f95c097e-87aa-4cf6-b1f9-a606463c406e\") " pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.652615 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb9wm\" (UniqueName: \"kubernetes.io/projected/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-kube-api-access-mb9wm\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.654727 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-serving-cert\") pod \"route-controller-manager-75fbfcbdd9-9h7cn\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.835572 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:01 crc kubenswrapper[4675]: I0219 08:49:01.846855 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:02 crc kubenswrapper[4675]: I0219 08:49:02.120817 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54f57c75f8-cnbng"] Feb 19 08:49:02 crc kubenswrapper[4675]: I0219 08:49:02.126360 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn"] Feb 19 08:49:02 crc kubenswrapper[4675]: W0219 08:49:02.130811 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf95c097e_87aa_4cf6_b1f9_a606463c406e.slice/crio-a0a20228478f2ed966f58857bb230b9f75655bdba195a12a342218ab4333db8d WatchSource:0}: Error finding container a0a20228478f2ed966f58857bb230b9f75655bdba195a12a342218ab4333db8d: Status 404 returned error can't find the container with id a0a20228478f2ed966f58857bb230b9f75655bdba195a12a342218ab4333db8d Feb 19 08:49:02 crc kubenswrapper[4675]: W0219 08:49:02.132749 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod189bce47_73c3_4f3a_a7e2_7e28b3fd5348.slice/crio-d307c666df42672bc828078280551d18ac930531272b205baba73bf838781781 WatchSource:0}: Error finding container d307c666df42672bc828078280551d18ac930531272b205baba73bf838781781: Status 404 returned error can't find the container with id d307c666df42672bc828078280551d18ac930531272b205baba73bf838781781 Feb 19 08:49:02 crc kubenswrapper[4675]: I0219 08:49:02.251870 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" event={"ID":"f95c097e-87aa-4cf6-b1f9-a606463c406e","Type":"ContainerStarted","Data":"a0a20228478f2ed966f58857bb230b9f75655bdba195a12a342218ab4333db8d"} Feb 19 08:49:02 crc kubenswrapper[4675]: I0219 08:49:02.253785 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" event={"ID":"189bce47-73c3-4f3a-a7e2-7e28b3fd5348","Type":"ContainerStarted","Data":"d307c666df42672bc828078280551d18ac930531272b205baba73bf838781781"} Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.273112 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" event={"ID":"f95c097e-87aa-4cf6-b1f9-a606463c406e","Type":"ContainerStarted","Data":"4edf73c773be715eef841dc649fd69d28609c754b8f76940c6d6d6530afbde08"} Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.273979 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.278595 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" event={"ID":"189bce47-73c3-4f3a-a7e2-7e28b3fd5348","Type":"ContainerStarted","Data":"d4efe8cce9f7ea509d60a614b2fa08d5d63f345dc7b64916da04f44a5452d803"} Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.278964 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.283545 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.284127 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.290706 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-54f57c75f8-cnbng" podStartSLOduration=4.290684739 podStartE2EDuration="4.290684739s" podCreationTimestamp="2026-02-19 08:48:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:49:03.288049613 +0000 UTC m=+324.915139881" watchObservedRunningTime="2026-02-19 08:49:03.290684739 +0000 UTC m=+324.917775017" Feb 19 08:49:03 crc kubenswrapper[4675]: I0219 08:49:03.336300 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" podStartSLOduration=4.336281896 podStartE2EDuration="4.336281896s" podCreationTimestamp="2026-02-19 08:48:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:49:03.332853167 +0000 UTC m=+324.959943435" watchObservedRunningTime="2026-02-19 08:49:03.336281896 +0000 UTC m=+324.963372164" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.223014 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn"] Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.224108 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" podUID="189bce47-73c3-4f3a-a7e2-7e28b3fd5348" containerName="route-controller-manager" containerID="cri-o://d4efe8cce9f7ea509d60a614b2fa08d5d63f345dc7b64916da04f44a5452d803" gracePeriod=30 Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.383094 4675 generic.go:334] "Generic (PLEG): container finished" podID="189bce47-73c3-4f3a-a7e2-7e28b3fd5348" containerID="d4efe8cce9f7ea509d60a614b2fa08d5d63f345dc7b64916da04f44a5452d803" exitCode=0 Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.383157 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" event={"ID":"189bce47-73c3-4f3a-a7e2-7e28b3fd5348","Type":"ContainerDied","Data":"d4efe8cce9f7ea509d60a614b2fa08d5d63f345dc7b64916da04f44a5452d803"} Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.681369 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.832317 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-serving-cert\") pod \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.832390 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb9wm\" (UniqueName: \"kubernetes.io/projected/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-kube-api-access-mb9wm\") pod \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.832492 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-config\") pod \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.832510 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-client-ca\") pod \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\" (UID: \"189bce47-73c3-4f3a-a7e2-7e28b3fd5348\") " Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.833182 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-client-ca" (OuterVolumeSpecName: "client-ca") pod "189bce47-73c3-4f3a-a7e2-7e28b3fd5348" (UID: "189bce47-73c3-4f3a-a7e2-7e28b3fd5348"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.833219 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-config" (OuterVolumeSpecName: "config") pod "189bce47-73c3-4f3a-a7e2-7e28b3fd5348" (UID: "189bce47-73c3-4f3a-a7e2-7e28b3fd5348"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.833289 4675 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-client-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.838920 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "189bce47-73c3-4f3a-a7e2-7e28b3fd5348" (UID: "189bce47-73c3-4f3a-a7e2-7e28b3fd5348"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.839433 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-kube-api-access-mb9wm" (OuterVolumeSpecName: "kube-api-access-mb9wm") pod "189bce47-73c3-4f3a-a7e2-7e28b3fd5348" (UID: "189bce47-73c3-4f3a-a7e2-7e28b3fd5348"). InnerVolumeSpecName "kube-api-access-mb9wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.934388 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.934815 4675 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:22 crc kubenswrapper[4675]: I0219 08:49:22.934942 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb9wm\" (UniqueName: \"kubernetes.io/projected/189bce47-73c3-4f3a-a7e2-7e28b3fd5348-kube-api-access-mb9wm\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.393680 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" event={"ID":"189bce47-73c3-4f3a-a7e2-7e28b3fd5348","Type":"ContainerDied","Data":"d307c666df42672bc828078280551d18ac930531272b205baba73bf838781781"} Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.393761 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.393768 4675 scope.go:117] "RemoveContainer" containerID="d4efe8cce9f7ea509d60a614b2fa08d5d63f345dc7b64916da04f44a5452d803" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.416155 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn"] Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.419285 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75fbfcbdd9-9h7cn"] Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.523445 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh"] Feb 19 08:49:23 crc kubenswrapper[4675]: E0219 08:49:23.523719 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189bce47-73c3-4f3a-a7e2-7e28b3fd5348" containerName="route-controller-manager" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.523738 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="189bce47-73c3-4f3a-a7e2-7e28b3fd5348" containerName="route-controller-manager" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.523860 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="189bce47-73c3-4f3a-a7e2-7e28b3fd5348" containerName="route-controller-manager" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.524490 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.527658 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.527753 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.527976 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.528082 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.530775 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.531532 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.534723 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh"] Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.646044 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da320000-5f15-4919-b1ab-2932d5e73ffb-client-ca\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.646461 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da320000-5f15-4919-b1ab-2932d5e73ffb-serving-cert\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.646572 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da320000-5f15-4919-b1ab-2932d5e73ffb-config\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.646692 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9scrx\" (UniqueName: \"kubernetes.io/projected/da320000-5f15-4919-b1ab-2932d5e73ffb-kube-api-access-9scrx\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.747892 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9scrx\" (UniqueName: \"kubernetes.io/projected/da320000-5f15-4919-b1ab-2932d5e73ffb-kube-api-access-9scrx\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.748349 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da320000-5f15-4919-b1ab-2932d5e73ffb-client-ca\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.748454 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da320000-5f15-4919-b1ab-2932d5e73ffb-config\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.748548 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da320000-5f15-4919-b1ab-2932d5e73ffb-serving-cert\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.749663 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da320000-5f15-4919-b1ab-2932d5e73ffb-client-ca\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.750037 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da320000-5f15-4919-b1ab-2932d5e73ffb-config\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.753734 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da320000-5f15-4919-b1ab-2932d5e73ffb-serving-cert\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.762571 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9scrx\" (UniqueName: \"kubernetes.io/projected/da320000-5f15-4919-b1ab-2932d5e73ffb-kube-api-access-9scrx\") pod \"route-controller-manager-6956b97674-tbxkh\" (UID: \"da320000-5f15-4919-b1ab-2932d5e73ffb\") " pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:23 crc kubenswrapper[4675]: I0219 08:49:23.849530 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:24 crc kubenswrapper[4675]: I0219 08:49:24.262332 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh"] Feb 19 08:49:24 crc kubenswrapper[4675]: I0219 08:49:24.399576 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" event={"ID":"da320000-5f15-4919-b1ab-2932d5e73ffb","Type":"ContainerStarted","Data":"ccaf852f35808f0a82d487f767f0379be71f234af55cb50247eb866e9d87f847"} Feb 19 08:49:24 crc kubenswrapper[4675]: I0219 08:49:24.400109 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" event={"ID":"da320000-5f15-4919-b1ab-2932d5e73ffb","Type":"ContainerStarted","Data":"eb614efbadf326148a1c45c2142f71cd1d3f7f8bddb7ba3e1337b69a7974d983"} Feb 19 08:49:24 crc kubenswrapper[4675]: I0219 08:49:24.400130 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:24 crc kubenswrapper[4675]: I0219 08:49:24.401302 4675 patch_prober.go:28] interesting pod/route-controller-manager-6956b97674-tbxkh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Feb 19 08:49:24 crc kubenswrapper[4675]: I0219 08:49:24.401350 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" podUID="da320000-5f15-4919-b1ab-2932d5e73ffb" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Feb 19 08:49:24 crc kubenswrapper[4675]: I0219 08:49:24.416297 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" podStartSLOduration=2.4162781669999998 podStartE2EDuration="2.416278167s" podCreationTimestamp="2026-02-19 08:49:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:49:24.413743106 +0000 UTC m=+346.040833374" watchObservedRunningTime="2026-02-19 08:49:24.416278167 +0000 UTC m=+346.043368435" Feb 19 08:49:25 crc kubenswrapper[4675]: I0219 08:49:25.110813 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="189bce47-73c3-4f3a-a7e2-7e28b3fd5348" path="/var/lib/kubelet/pods/189bce47-73c3-4f3a-a7e2-7e28b3fd5348/volumes" Feb 19 08:49:25 crc kubenswrapper[4675]: I0219 08:49:25.409588 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6956b97674-tbxkh" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.674434 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbjkd"] Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.675444 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vbjkd" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="registry-server" containerID="cri-o://7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" gracePeriod=30 Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.686117 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vc6hh"] Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.686508 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vc6hh" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="registry-server" containerID="cri-o://245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" gracePeriod=30 Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.692314 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cwrvv"] Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.692614 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" containerID="cri-o://e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862" gracePeriod=30 Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.696450 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgxzf"] Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.696730 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sgxzf" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="registry-server" containerID="cri-o://66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd" gracePeriod=30 Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.714675 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6mmqj"] Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.714935 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6mmqj" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="registry-server" containerID="cri-o://59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6" gracePeriod=30 Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.717855 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c4g5q"] Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.718723 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.723307 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c4g5q"] Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.781152 4675 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9 is running failed: container process not found" containerID="245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" cmd=["grpc_health_probe","-addr=:50051"] Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.782192 4675 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9 is running failed: container process not found" containerID="245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" cmd=["grpc_health_probe","-addr=:50051"] Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.782563 4675 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9 is running failed: container process not found" containerID="245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" cmd=["grpc_health_probe","-addr=:50051"] Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.782751 4675 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-vc6hh" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="registry-server" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.862675 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/147346d6-fe78-404a-964e-4ee12c505b82-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.862729 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75v7h\" (UniqueName: \"kubernetes.io/projected/147346d6-fe78-404a-964e-4ee12c505b82-kube-api-access-75v7h\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.862839 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/147346d6-fe78-404a-964e-4ee12c505b82-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.964178 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/147346d6-fe78-404a-964e-4ee12c505b82-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.964252 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/147346d6-fe78-404a-964e-4ee12c505b82-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.964278 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75v7h\" (UniqueName: \"kubernetes.io/projected/147346d6-fe78-404a-964e-4ee12c505b82-kube-api-access-75v7h\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.965662 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/147346d6-fe78-404a-964e-4ee12c505b82-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.974352 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/147346d6-fe78-404a-964e-4ee12c505b82-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: I0219 08:49:38.980620 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75v7h\" (UniqueName: \"kubernetes.io/projected/147346d6-fe78-404a-964e-4ee12c505b82-kube-api-access-75v7h\") pod \"marketplace-operator-79b997595-c4g5q\" (UID: \"147346d6-fe78-404a-964e-4ee12c505b82\") " pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.987867 4675 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700 is running failed: container process not found" containerID="7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" cmd=["grpc_health_probe","-addr=:50051"] Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.988590 4675 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700 is running failed: container process not found" containerID="7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" cmd=["grpc_health_probe","-addr=:50051"] Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.988951 4675 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700 is running failed: container process not found" containerID="7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" cmd=["grpc_health_probe","-addr=:50051"] Feb 19 08:49:38 crc kubenswrapper[4675]: E0219 08:49:38.989037 4675 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-vbjkd" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="registry-server" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.052974 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.185943 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.267100 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz852\" (UniqueName: \"kubernetes.io/projected/52c2e6ec-6246-4346-8d44-f9d37b805bb0-kube-api-access-xz852\") pod \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.267150 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-utilities\") pod \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.267201 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-catalog-content\") pod \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\" (UID: \"52c2e6ec-6246-4346-8d44-f9d37b805bb0\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.271913 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-utilities" (OuterVolumeSpecName: "utilities") pod "52c2e6ec-6246-4346-8d44-f9d37b805bb0" (UID: "52c2e6ec-6246-4346-8d44-f9d37b805bb0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.276134 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c2e6ec-6246-4346-8d44-f9d37b805bb0-kube-api-access-xz852" (OuterVolumeSpecName: "kube-api-access-xz852") pod "52c2e6ec-6246-4346-8d44-f9d37b805bb0" (UID: "52c2e6ec-6246-4346-8d44-f9d37b805bb0"). InnerVolumeSpecName "kube-api-access-xz852". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.282917 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.293682 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.301418 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.324156 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.356755 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52c2e6ec-6246-4346-8d44-f9d37b805bb0" (UID: "52c2e6ec-6246-4346-8d44-f9d37b805bb0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.368827 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz852\" (UniqueName: \"kubernetes.io/projected/52c2e6ec-6246-4346-8d44-f9d37b805bb0-kube-api-access-xz852\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.368868 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.368883 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52c2e6ec-6246-4346-8d44-f9d37b805bb0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469303 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-catalog-content\") pod \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469362 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-trusted-ca\") pod \"017fb59c-7331-42bf-9a21-8dff74814740\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469391 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-operator-metrics\") pod \"017fb59c-7331-42bf-9a21-8dff74814740\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469419 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pskkw\" (UniqueName: \"kubernetes.io/projected/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-kube-api-access-pskkw\") pod \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469457 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgbbr\" (UniqueName: \"kubernetes.io/projected/017fb59c-7331-42bf-9a21-8dff74814740-kube-api-access-sgbbr\") pod \"017fb59c-7331-42bf-9a21-8dff74814740\" (UID: \"017fb59c-7331-42bf-9a21-8dff74814740\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469487 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-utilities\") pod \"b2ff348d-9b80-4358-afcf-dace55aef217\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469508 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlhr2\" (UniqueName: \"kubernetes.io/projected/60ff6ed9-e694-4050-aae5-f22f390208e7-kube-api-access-jlhr2\") pod \"60ff6ed9-e694-4050-aae5-f22f390208e7\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469538 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-utilities\") pod \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\" (UID: \"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469559 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-catalog-content\") pod \"b2ff348d-9b80-4358-afcf-dace55aef217\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469596 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-utilities\") pod \"60ff6ed9-e694-4050-aae5-f22f390208e7\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469630 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-catalog-content\") pod \"60ff6ed9-e694-4050-aae5-f22f390208e7\" (UID: \"60ff6ed9-e694-4050-aae5-f22f390208e7\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.469674 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7ch8\" (UniqueName: \"kubernetes.io/projected/b2ff348d-9b80-4358-afcf-dace55aef217-kube-api-access-p7ch8\") pod \"b2ff348d-9b80-4358-afcf-dace55aef217\" (UID: \"b2ff348d-9b80-4358-afcf-dace55aef217\") " Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.470232 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "017fb59c-7331-42bf-9a21-8dff74814740" (UID: "017fb59c-7331-42bf-9a21-8dff74814740"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.470760 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-utilities" (OuterVolumeSpecName: "utilities") pod "cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" (UID: "cfcc9c58-9792-449b-9fb8-faa5ef40cbc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.472085 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-utilities" (OuterVolumeSpecName: "utilities") pod "b2ff348d-9b80-4358-afcf-dace55aef217" (UID: "b2ff348d-9b80-4358-afcf-dace55aef217"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.472165 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-utilities" (OuterVolumeSpecName: "utilities") pod "60ff6ed9-e694-4050-aae5-f22f390208e7" (UID: "60ff6ed9-e694-4050-aae5-f22f390208e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.475491 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60ff6ed9-e694-4050-aae5-f22f390208e7-kube-api-access-jlhr2" (OuterVolumeSpecName: "kube-api-access-jlhr2") pod "60ff6ed9-e694-4050-aae5-f22f390208e7" (UID: "60ff6ed9-e694-4050-aae5-f22f390208e7"). InnerVolumeSpecName "kube-api-access-jlhr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.476323 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/017fb59c-7331-42bf-9a21-8dff74814740-kube-api-access-sgbbr" (OuterVolumeSpecName: "kube-api-access-sgbbr") pod "017fb59c-7331-42bf-9a21-8dff74814740" (UID: "017fb59c-7331-42bf-9a21-8dff74814740"). InnerVolumeSpecName "kube-api-access-sgbbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.476771 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ff348d-9b80-4358-afcf-dace55aef217-kube-api-access-p7ch8" (OuterVolumeSpecName: "kube-api-access-p7ch8") pod "b2ff348d-9b80-4358-afcf-dace55aef217" (UID: "b2ff348d-9b80-4358-afcf-dace55aef217"). InnerVolumeSpecName "kube-api-access-p7ch8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.477065 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "017fb59c-7331-42bf-9a21-8dff74814740" (UID: "017fb59c-7331-42bf-9a21-8dff74814740"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.479111 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-kube-api-access-pskkw" (OuterVolumeSpecName: "kube-api-access-pskkw") pod "cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" (UID: "cfcc9c58-9792-449b-9fb8-faa5ef40cbc2"). InnerVolumeSpecName "kube-api-access-pskkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.485123 4675 generic.go:334] "Generic (PLEG): container finished" podID="b2ff348d-9b80-4358-afcf-dace55aef217" containerID="66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd" exitCode=0 Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.485195 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgxzf" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.485264 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgxzf" event={"ID":"b2ff348d-9b80-4358-afcf-dace55aef217","Type":"ContainerDied","Data":"66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.485348 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgxzf" event={"ID":"b2ff348d-9b80-4358-afcf-dace55aef217","Type":"ContainerDied","Data":"1aec6ff40791d740027d8dce7864b4dd4b344a7fd1ed543160457ca5e0b32db8"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.485382 4675 scope.go:117] "RemoveContainer" containerID="66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.493019 4675 generic.go:334] "Generic (PLEG): container finished" podID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerID="7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" exitCode=0 Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.493119 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbjkd" event={"ID":"52c2e6ec-6246-4346-8d44-f9d37b805bb0","Type":"ContainerDied","Data":"7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.493156 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbjkd" event={"ID":"52c2e6ec-6246-4346-8d44-f9d37b805bb0","Type":"ContainerDied","Data":"e4917ffa058b7f8e2b701a48eb05eb9617ddcc71671745cf03b02c219403f5a1"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.493169 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbjkd" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.496309 4675 generic.go:334] "Generic (PLEG): container finished" podID="017fb59c-7331-42bf-9a21-8dff74814740" containerID="e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862" exitCode=0 Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.496354 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.496408 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" event={"ID":"017fb59c-7331-42bf-9a21-8dff74814740","Type":"ContainerDied","Data":"e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.496440 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cwrvv" event={"ID":"017fb59c-7331-42bf-9a21-8dff74814740","Type":"ContainerDied","Data":"59a9ee2a13ded4c16f09461c51fc18c8db80bbe4a494f8aeddc395e53cdebfda"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.502821 4675 generic.go:334] "Generic (PLEG): container finished" podID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerID="59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6" exitCode=0 Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.502894 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerDied","Data":"59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.502924 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mmqj" event={"ID":"60ff6ed9-e694-4050-aae5-f22f390208e7","Type":"ContainerDied","Data":"02486411a3a4ea0a7e60ca4a5774e94f083421cb95abdabafd20589fc0799d38"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.503272 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mmqj" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.507500 4675 generic.go:334] "Generic (PLEG): container finished" podID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerID="245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" exitCode=0 Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.507529 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc6hh" event={"ID":"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2","Type":"ContainerDied","Data":"245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.507547 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc6hh" event={"ID":"cfcc9c58-9792-449b-9fb8-faa5ef40cbc2","Type":"ContainerDied","Data":"69dca1d13c27bde045bd6c04f3c3862002065ed38d9514273ff7de814388c5a9"} Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.507595 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc6hh" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.507895 4675 scope.go:117] "RemoveContainer" containerID="79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.521653 4675 scope.go:117] "RemoveContainer" containerID="9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.523224 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" (UID: "cfcc9c58-9792-449b-9fb8-faa5ef40cbc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.544431 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cwrvv"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.549950 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2ff348d-9b80-4358-afcf-dace55aef217" (UID: "b2ff348d-9b80-4358-afcf-dace55aef217"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.558263 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cwrvv"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.559883 4675 scope.go:117] "RemoveContainer" containerID="66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.560391 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd\": container with ID starting with 66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd not found: ID does not exist" containerID="66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.560442 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd"} err="failed to get container status \"66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd\": rpc error: code = NotFound desc = could not find container \"66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd\": container with ID starting with 66750344e34dec53e94212b9f7afcbf4fe340b88f21d4afbdd43520598cfd0bd not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.560477 4675 scope.go:117] "RemoveContainer" containerID="79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.560793 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837\": container with ID starting with 79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837 not found: ID does not exist" containerID="79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.560822 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837"} err="failed to get container status \"79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837\": rpc error: code = NotFound desc = could not find container \"79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837\": container with ID starting with 79d29bc915fd46d0d55a5ce9db63a01864ab3574fd41091124f39c3158626837 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.560846 4675 scope.go:117] "RemoveContainer" containerID="9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.561416 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250\": container with ID starting with 9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250 not found: ID does not exist" containerID="9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.561446 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250"} err="failed to get container status \"9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250\": rpc error: code = NotFound desc = could not find container \"9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250\": container with ID starting with 9b8a6d963504fe61d46d02a7190098c476846984c9698031cd14d11a64c23250 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.561462 4675 scope.go:117] "RemoveContainer" containerID="7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.561999 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbjkd"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.566028 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vbjkd"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.570606 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c4g5q"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571384 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571416 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571428 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7ch8\" (UniqueName: \"kubernetes.io/projected/b2ff348d-9b80-4358-afcf-dace55aef217-kube-api-access-p7ch8\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571439 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571448 4675 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571457 4675 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/017fb59c-7331-42bf-9a21-8dff74814740-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571469 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pskkw\" (UniqueName: \"kubernetes.io/projected/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-kube-api-access-pskkw\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571479 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgbbr\" (UniqueName: \"kubernetes.io/projected/017fb59c-7331-42bf-9a21-8dff74814740-kube-api-access-sgbbr\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571490 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ff348d-9b80-4358-afcf-dace55aef217-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571498 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlhr2\" (UniqueName: \"kubernetes.io/projected/60ff6ed9-e694-4050-aae5-f22f390208e7-kube-api-access-jlhr2\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.571506 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.584963 4675 scope.go:117] "RemoveContainer" containerID="b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5" Feb 19 08:49:39 crc kubenswrapper[4675]: W0219 08:49:39.588538 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod147346d6_fe78_404a_964e_4ee12c505b82.slice/crio-5bd02dd8fcc9301c37bf5e893fb248417be592d75c4d56bde623eda6dbccb1ee WatchSource:0}: Error finding container 5bd02dd8fcc9301c37bf5e893fb248417be592d75c4d56bde623eda6dbccb1ee: Status 404 returned error can't find the container with id 5bd02dd8fcc9301c37bf5e893fb248417be592d75c4d56bde623eda6dbccb1ee Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.609145 4675 scope.go:117] "RemoveContainer" containerID="4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.630585 4675 scope.go:117] "RemoveContainer" containerID="7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.631074 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700\": container with ID starting with 7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700 not found: ID does not exist" containerID="7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.631112 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700"} err="failed to get container status \"7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700\": rpc error: code = NotFound desc = could not find container \"7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700\": container with ID starting with 7fee3279f75d5a95a1e156c3213167340b211553ff358682c191719f8884a700 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.631138 4675 scope.go:117] "RemoveContainer" containerID="b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.631593 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5\": container with ID starting with b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5 not found: ID does not exist" containerID="b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.631658 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5"} err="failed to get container status \"b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5\": rpc error: code = NotFound desc = could not find container \"b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5\": container with ID starting with b5290ba19019d4ba19f77ebe73a3545061b0ebf9bcc4f91240df256306e112c5 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.631694 4675 scope.go:117] "RemoveContainer" containerID="4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.632068 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d\": container with ID starting with 4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d not found: ID does not exist" containerID="4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.632120 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d"} err="failed to get container status \"4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d\": rpc error: code = NotFound desc = could not find container \"4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d\": container with ID starting with 4522556ac3ad0e2c62b7c738a7c9489438a7f86118cbe0ecc5380664fddb124d not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.632133 4675 scope.go:117] "RemoveContainer" containerID="e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.645955 4675 scope.go:117] "RemoveContainer" containerID="890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.651801 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60ff6ed9-e694-4050-aae5-f22f390208e7" (UID: "60ff6ed9-e694-4050-aae5-f22f390208e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.665095 4675 scope.go:117] "RemoveContainer" containerID="e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.665572 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862\": container with ID starting with e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862 not found: ID does not exist" containerID="e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.665603 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862"} err="failed to get container status \"e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862\": rpc error: code = NotFound desc = could not find container \"e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862\": container with ID starting with e44e663fb349dd5e1afff053d81d0841d946b7d29e9b8b33e5f4486164814862 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.665627 4675 scope.go:117] "RemoveContainer" containerID="890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.666031 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8\": container with ID starting with 890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8 not found: ID does not exist" containerID="890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.666103 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8"} err="failed to get container status \"890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8\": rpc error: code = NotFound desc = could not find container \"890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8\": container with ID starting with 890f1babd75bc2f50b10baad60a0fb48667d964d960e9edfc8dd4247b17afbc8 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.666168 4675 scope.go:117] "RemoveContainer" containerID="59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.672624 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60ff6ed9-e694-4050-aae5-f22f390208e7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.680223 4675 scope.go:117] "RemoveContainer" containerID="00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.735927 4675 scope.go:117] "RemoveContainer" containerID="8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.751094 4675 scope.go:117] "RemoveContainer" containerID="59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.751551 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6\": container with ID starting with 59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6 not found: ID does not exist" containerID="59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.751600 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6"} err="failed to get container status \"59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6\": rpc error: code = NotFound desc = could not find container \"59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6\": container with ID starting with 59fe6be43faca5bc4392efcee0a4dce49291ac6730b1052f56e1c0447ee7b9e6 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.751684 4675 scope.go:117] "RemoveContainer" containerID="00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.752066 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876\": container with ID starting with 00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876 not found: ID does not exist" containerID="00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.752096 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876"} err="failed to get container status \"00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876\": rpc error: code = NotFound desc = could not find container \"00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876\": container with ID starting with 00db668b596c4a926050e2c9d586a3842b0b4a95baeff8af715183298fd1c876 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.752119 4675 scope.go:117] "RemoveContainer" containerID="8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.752614 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d\": container with ID starting with 8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d not found: ID does not exist" containerID="8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.752633 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d"} err="failed to get container status \"8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d\": rpc error: code = NotFound desc = could not find container \"8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d\": container with ID starting with 8a8a291c30977ce193f8248c8d534977eab3ad8ca0e4bef99c75c292886ced3d not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.752659 4675 scope.go:117] "RemoveContainer" containerID="245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.764593 4675 scope.go:117] "RemoveContainer" containerID="80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.783087 4675 scope.go:117] "RemoveContainer" containerID="60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.819096 4675 scope.go:117] "RemoveContainer" containerID="245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.825306 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9\": container with ID starting with 245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9 not found: ID does not exist" containerID="245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.825380 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9"} err="failed to get container status \"245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9\": rpc error: code = NotFound desc = could not find container \"245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9\": container with ID starting with 245ef48b89c4cddf98ab07b4aa293c582fd219e66acb34d3ef5016044d5ea8a9 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.825430 4675 scope.go:117] "RemoveContainer" containerID="80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.826005 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e\": container with ID starting with 80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e not found: ID does not exist" containerID="80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.826067 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e"} err="failed to get container status \"80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e\": rpc error: code = NotFound desc = could not find container \"80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e\": container with ID starting with 80cf1be969401e495f1c6021bd3b1d4296c58c82a54a245948ac8781a8082b8e not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.826113 4675 scope.go:117] "RemoveContainer" containerID="60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3" Feb 19 08:49:39 crc kubenswrapper[4675]: E0219 08:49:39.826385 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3\": container with ID starting with 60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3 not found: ID does not exist" containerID="60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.826409 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3"} err="failed to get container status \"60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3\": rpc error: code = NotFound desc = could not find container \"60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3\": container with ID starting with 60303fa328acf075d4daaa6b95cf62b8688e6ec4e6032dd6e13b8aecb403ffd3 not found: ID does not exist" Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.830115 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgxzf"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.834986 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgxzf"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.853240 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vc6hh"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.861435 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vc6hh"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.864527 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6mmqj"] Feb 19 08:49:39 crc kubenswrapper[4675]: I0219 08:49:39.867587 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6mmqj"] Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.515032 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" event={"ID":"147346d6-fe78-404a-964e-4ee12c505b82","Type":"ContainerStarted","Data":"992f10d694e54d1b28956e6e4543e9c00ac64b2b6dc71690994a5878b5177349"} Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.515097 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" event={"ID":"147346d6-fe78-404a-964e-4ee12c505b82","Type":"ContainerStarted","Data":"5bd02dd8fcc9301c37bf5e893fb248417be592d75c4d56bde623eda6dbccb1ee"} Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.515295 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.518431 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.534435 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-c4g5q" podStartSLOduration=2.534414801 podStartE2EDuration="2.534414801s" podCreationTimestamp="2026-02-19 08:49:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:49:40.532757004 +0000 UTC m=+362.159847272" watchObservedRunningTime="2026-02-19 08:49:40.534414801 +0000 UTC m=+362.161505069" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805461 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fg4v6"] Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805711 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805731 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805748 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805756 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805766 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805773 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805790 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805799 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805810 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805816 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805828 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805836 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805849 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805856 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805863 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805871 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805879 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805886 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805900 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805907 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805917 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805926 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="extract-content" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805937 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805944 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="extract-utilities" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805956 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805964 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: E0219 08:49:40.805972 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.805979 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.806087 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.806104 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.806124 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.806133 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.806144 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" containerName="registry-server" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.806338 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="017fb59c-7331-42bf-9a21-8dff74814740" containerName="marketplace-operator" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.807010 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.808952 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.815430 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg4v6"] Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.898296 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh7hq\" (UniqueName: \"kubernetes.io/projected/3d9580d3-8d3f-4730-af25-6c57eead5415-kube-api-access-jh7hq\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.898356 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9580d3-8d3f-4730-af25-6c57eead5415-utilities\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:40 crc kubenswrapper[4675]: I0219 08:49:40.898489 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9580d3-8d3f-4730-af25-6c57eead5415-catalog-content\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:40.999837 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9580d3-8d3f-4730-af25-6c57eead5415-catalog-content\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.000361 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh7hq\" (UniqueName: \"kubernetes.io/projected/3d9580d3-8d3f-4730-af25-6c57eead5415-kube-api-access-jh7hq\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.000403 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9580d3-8d3f-4730-af25-6c57eead5415-utilities\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.000661 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9580d3-8d3f-4730-af25-6c57eead5415-catalog-content\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.005251 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9580d3-8d3f-4730-af25-6c57eead5415-utilities\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.007691 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8xtr4"] Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.008754 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.011304 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.026417 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xtr4"] Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.029403 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh7hq\" (UniqueName: \"kubernetes.io/projected/3d9580d3-8d3f-4730-af25-6c57eead5415-kube-api-access-jh7hq\") pod \"redhat-marketplace-fg4v6\" (UID: \"3d9580d3-8d3f-4730-af25-6c57eead5415\") " pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.102144 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-754cg\" (UniqueName: \"kubernetes.io/projected/b2b9392d-292a-40aa-9271-805bde5fd1d6-kube-api-access-754cg\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.102199 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b9392d-292a-40aa-9271-805bde5fd1d6-utilities\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.102254 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b9392d-292a-40aa-9271-805bde5fd1d6-catalog-content\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.109386 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="017fb59c-7331-42bf-9a21-8dff74814740" path="/var/lib/kubelet/pods/017fb59c-7331-42bf-9a21-8dff74814740/volumes" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.109877 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52c2e6ec-6246-4346-8d44-f9d37b805bb0" path="/var/lib/kubelet/pods/52c2e6ec-6246-4346-8d44-f9d37b805bb0/volumes" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.110574 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60ff6ed9-e694-4050-aae5-f22f390208e7" path="/var/lib/kubelet/pods/60ff6ed9-e694-4050-aae5-f22f390208e7/volumes" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.111693 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ff348d-9b80-4358-afcf-dace55aef217" path="/var/lib/kubelet/pods/b2ff348d-9b80-4358-afcf-dace55aef217/volumes" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.112247 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfcc9c58-9792-449b-9fb8-faa5ef40cbc2" path="/var/lib/kubelet/pods/cfcc9c58-9792-449b-9fb8-faa5ef40cbc2/volumes" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.165360 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.203727 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-754cg\" (UniqueName: \"kubernetes.io/projected/b2b9392d-292a-40aa-9271-805bde5fd1d6-kube-api-access-754cg\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.203787 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b9392d-292a-40aa-9271-805bde5fd1d6-utilities\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.203810 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b9392d-292a-40aa-9271-805bde5fd1d6-catalog-content\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.204400 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b9392d-292a-40aa-9271-805bde5fd1d6-utilities\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.205704 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b9392d-292a-40aa-9271-805bde5fd1d6-catalog-content\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.226477 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-754cg\" (UniqueName: \"kubernetes.io/projected/b2b9392d-292a-40aa-9271-805bde5fd1d6-kube-api-access-754cg\") pod \"redhat-operators-8xtr4\" (UID: \"b2b9392d-292a-40aa-9271-805bde5fd1d6\") " pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.329004 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.562860 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fg4v6"] Feb 19 08:49:41 crc kubenswrapper[4675]: W0219 08:49:41.568930 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d9580d3_8d3f_4730_af25_6c57eead5415.slice/crio-edcaa5f73f5f50c39e8d95a6ebc667a1c4516dbea720793da33c82716b366f4b WatchSource:0}: Error finding container edcaa5f73f5f50c39e8d95a6ebc667a1c4516dbea720793da33c82716b366f4b: Status 404 returned error can't find the container with id edcaa5f73f5f50c39e8d95a6ebc667a1c4516dbea720793da33c82716b366f4b Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.707315 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8xtr4"] Feb 19 08:49:41 crc kubenswrapper[4675]: W0219 08:49:41.713694 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2b9392d_292a_40aa_9271_805bde5fd1d6.slice/crio-73af064d07a1cae450df61f46b574571c6866e092fbd40bf0120892a7bc1ad69 WatchSource:0}: Error finding container 73af064d07a1cae450df61f46b574571c6866e092fbd40bf0120892a7bc1ad69: Status 404 returned error can't find the container with id 73af064d07a1cae450df61f46b574571c6866e092fbd40bf0120892a7bc1ad69 Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.743385 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:49:41 crc kubenswrapper[4675]: I0219 08:49:41.743473 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:49:42 crc kubenswrapper[4675]: I0219 08:49:42.533559 4675 generic.go:334] "Generic (PLEG): container finished" podID="b2b9392d-292a-40aa-9271-805bde5fd1d6" containerID="5ba50b7cc830c7fd4cc6fb4bde9ead92b5bfcc9b13364191738d79333e749ecb" exitCode=0 Feb 19 08:49:42 crc kubenswrapper[4675]: I0219 08:49:42.534083 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xtr4" event={"ID":"b2b9392d-292a-40aa-9271-805bde5fd1d6","Type":"ContainerDied","Data":"5ba50b7cc830c7fd4cc6fb4bde9ead92b5bfcc9b13364191738d79333e749ecb"} Feb 19 08:49:42 crc kubenswrapper[4675]: I0219 08:49:42.534119 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xtr4" event={"ID":"b2b9392d-292a-40aa-9271-805bde5fd1d6","Type":"ContainerStarted","Data":"73af064d07a1cae450df61f46b574571c6866e092fbd40bf0120892a7bc1ad69"} Feb 19 08:49:42 crc kubenswrapper[4675]: I0219 08:49:42.537098 4675 generic.go:334] "Generic (PLEG): container finished" podID="3d9580d3-8d3f-4730-af25-6c57eead5415" containerID="d1127c5a050625285e3c842db4a68b03f75a5bd919ec7c942f641dc50888dea8" exitCode=0 Feb 19 08:49:42 crc kubenswrapper[4675]: I0219 08:49:42.537135 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg4v6" event={"ID":"3d9580d3-8d3f-4730-af25-6c57eead5415","Type":"ContainerDied","Data":"d1127c5a050625285e3c842db4a68b03f75a5bd919ec7c942f641dc50888dea8"} Feb 19 08:49:42 crc kubenswrapper[4675]: I0219 08:49:42.537181 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg4v6" event={"ID":"3d9580d3-8d3f-4730-af25-6c57eead5415","Type":"ContainerStarted","Data":"edcaa5f73f5f50c39e8d95a6ebc667a1c4516dbea720793da33c82716b366f4b"} Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.412295 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gmkvv"] Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.413629 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.416800 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.421959 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmkvv"] Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.430538 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-utilities\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.430624 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-catalog-content\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.430664 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcdgm\" (UniqueName: \"kubernetes.io/projected/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-kube-api-access-gcdgm\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.531890 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-catalog-content\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.531945 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcdgm\" (UniqueName: \"kubernetes.io/projected/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-kube-api-access-gcdgm\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.532026 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-utilities\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.532430 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-utilities\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.532682 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-catalog-content\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.553730 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcdgm\" (UniqueName: \"kubernetes.io/projected/4adf2b0e-ea04-47fb-8c06-e8f2902d1257-kube-api-access-gcdgm\") pod \"community-operators-gmkvv\" (UID: \"4adf2b0e-ea04-47fb-8c06-e8f2902d1257\") " pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.563388 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xtr4" event={"ID":"b2b9392d-292a-40aa-9271-805bde5fd1d6","Type":"ContainerStarted","Data":"13513e80501b1385a9e9dc7915720875e2650b1f4ea7d0512239ac0fbbe0c2b5"} Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.565536 4675 generic.go:334] "Generic (PLEG): container finished" podID="3d9580d3-8d3f-4730-af25-6c57eead5415" containerID="ae8e598649dfe64645380c73b39ff9055b4f6d2b305124f1ba9d9f435059d013" exitCode=0 Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.565580 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg4v6" event={"ID":"3d9580d3-8d3f-4730-af25-6c57eead5415","Type":"ContainerDied","Data":"ae8e598649dfe64645380c73b39ff9055b4f6d2b305124f1ba9d9f435059d013"} Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.611451 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mvrwn"] Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.612609 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.619541 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.633026 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsmjt\" (UniqueName: \"kubernetes.io/projected/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-kube-api-access-vsmjt\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.633108 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-catalog-content\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.633194 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-utilities\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.666338 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mvrwn"] Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.734463 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsmjt\" (UniqueName: \"kubernetes.io/projected/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-kube-api-access-vsmjt\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.734524 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-catalog-content\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.734577 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-utilities\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.735045 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-utilities\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.735546 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-catalog-content\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.738135 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.751793 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsmjt\" (UniqueName: \"kubernetes.io/projected/50ddb4cd-7719-4f7a-a71f-a8bd832236e0-kube-api-access-vsmjt\") pod \"certified-operators-mvrwn\" (UID: \"50ddb4cd-7719-4f7a-a71f-a8bd832236e0\") " pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.930198 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:43 crc kubenswrapper[4675]: I0219 08:49:43.941234 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmkvv"] Feb 19 08:49:43 crc kubenswrapper[4675]: W0219 08:49:43.957066 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4adf2b0e_ea04_47fb_8c06_e8f2902d1257.slice/crio-01b32b7ec6c69ecf2587345a4292fa68a64013a8a26b1274e5cc5ef1ad516932 WatchSource:0}: Error finding container 01b32b7ec6c69ecf2587345a4292fa68a64013a8a26b1274e5cc5ef1ad516932: Status 404 returned error can't find the container with id 01b32b7ec6c69ecf2587345a4292fa68a64013a8a26b1274e5cc5ef1ad516932 Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.169083 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mvrwn"] Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.573609 4675 generic.go:334] "Generic (PLEG): container finished" podID="4adf2b0e-ea04-47fb-8c06-e8f2902d1257" containerID="164cef627f687ab492e57eef1b0451727ee0cefe40c81959847f42461819b427" exitCode=0 Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.573690 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmkvv" event={"ID":"4adf2b0e-ea04-47fb-8c06-e8f2902d1257","Type":"ContainerDied","Data":"164cef627f687ab492e57eef1b0451727ee0cefe40c81959847f42461819b427"} Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.574154 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmkvv" event={"ID":"4adf2b0e-ea04-47fb-8c06-e8f2902d1257","Type":"ContainerStarted","Data":"01b32b7ec6c69ecf2587345a4292fa68a64013a8a26b1274e5cc5ef1ad516932"} Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.577168 4675 generic.go:334] "Generic (PLEG): container finished" podID="b2b9392d-292a-40aa-9271-805bde5fd1d6" containerID="13513e80501b1385a9e9dc7915720875e2650b1f4ea7d0512239ac0fbbe0c2b5" exitCode=0 Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.577241 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xtr4" event={"ID":"b2b9392d-292a-40aa-9271-805bde5fd1d6","Type":"ContainerDied","Data":"13513e80501b1385a9e9dc7915720875e2650b1f4ea7d0512239ac0fbbe0c2b5"} Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.579797 4675 generic.go:334] "Generic (PLEG): container finished" podID="50ddb4cd-7719-4f7a-a71f-a8bd832236e0" containerID="6274d21cb00823a7aa6b548249520b76e6dad20997209c664ae6712505cb7608" exitCode=0 Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.580473 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrwn" event={"ID":"50ddb4cd-7719-4f7a-a71f-a8bd832236e0","Type":"ContainerDied","Data":"6274d21cb00823a7aa6b548249520b76e6dad20997209c664ae6712505cb7608"} Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.580513 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrwn" event={"ID":"50ddb4cd-7719-4f7a-a71f-a8bd832236e0","Type":"ContainerStarted","Data":"28530c0db107fb750386083b71b935eb755bc75ec1dff32dadb7a0f9ff67f6dd"} Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.597152 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fg4v6" event={"ID":"3d9580d3-8d3f-4730-af25-6c57eead5415","Type":"ContainerStarted","Data":"9a634a96921165e598e9898e3a0f355e6c318fb44fb270c90496d60b9c670476"} Feb 19 08:49:44 crc kubenswrapper[4675]: I0219 08:49:44.626472 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fg4v6" podStartSLOduration=3.160789401 podStartE2EDuration="4.626447678s" podCreationTimestamp="2026-02-19 08:49:40 +0000 UTC" firstStartedPulling="2026-02-19 08:49:42.538920517 +0000 UTC m=+364.166010785" lastFinishedPulling="2026-02-19 08:49:44.004578794 +0000 UTC m=+365.631669062" observedRunningTime="2026-02-19 08:49:44.624624897 +0000 UTC m=+366.251715165" watchObservedRunningTime="2026-02-19 08:49:44.626447678 +0000 UTC m=+366.253537946" Feb 19 08:49:45 crc kubenswrapper[4675]: I0219 08:49:45.607024 4675 generic.go:334] "Generic (PLEG): container finished" podID="4adf2b0e-ea04-47fb-8c06-e8f2902d1257" containerID="484351d2645a3f95de03676691b1be2f896876c75c10733ecf5d6ff86e5f10be" exitCode=0 Feb 19 08:49:45 crc kubenswrapper[4675]: I0219 08:49:45.607139 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmkvv" event={"ID":"4adf2b0e-ea04-47fb-8c06-e8f2902d1257","Type":"ContainerDied","Data":"484351d2645a3f95de03676691b1be2f896876c75c10733ecf5d6ff86e5f10be"} Feb 19 08:49:45 crc kubenswrapper[4675]: I0219 08:49:45.610482 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8xtr4" event={"ID":"b2b9392d-292a-40aa-9271-805bde5fd1d6","Type":"ContainerStarted","Data":"db94f6cfb12de4264328db48d1dd1a1bad31cc0ac7cd39ce4aa9f4090b54c696"} Feb 19 08:49:45 crc kubenswrapper[4675]: I0219 08:49:45.613362 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrwn" event={"ID":"50ddb4cd-7719-4f7a-a71f-a8bd832236e0","Type":"ContainerStarted","Data":"662cb6923858d2161f5e5090d0b7444133ecb02fa7705c65598b4204c64a4f08"} Feb 19 08:49:45 crc kubenswrapper[4675]: I0219 08:49:45.640597 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8xtr4" podStartSLOduration=3.207412719 podStartE2EDuration="5.64057533s" podCreationTimestamp="2026-02-19 08:49:40 +0000 UTC" firstStartedPulling="2026-02-19 08:49:42.536335315 +0000 UTC m=+364.163425583" lastFinishedPulling="2026-02-19 08:49:44.969497926 +0000 UTC m=+366.596588194" observedRunningTime="2026-02-19 08:49:45.640362654 +0000 UTC m=+367.267452922" watchObservedRunningTime="2026-02-19 08:49:45.64057533 +0000 UTC m=+367.267665598" Feb 19 08:49:46 crc kubenswrapper[4675]: I0219 08:49:46.625325 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmkvv" event={"ID":"4adf2b0e-ea04-47fb-8c06-e8f2902d1257","Type":"ContainerStarted","Data":"5a7ee89436a8d9d6d279632e9f914417cddb74d9e7b5306978635bcbcd4ca592"} Feb 19 08:49:46 crc kubenswrapper[4675]: I0219 08:49:46.627240 4675 generic.go:334] "Generic (PLEG): container finished" podID="50ddb4cd-7719-4f7a-a71f-a8bd832236e0" containerID="662cb6923858d2161f5e5090d0b7444133ecb02fa7705c65598b4204c64a4f08" exitCode=0 Feb 19 08:49:46 crc kubenswrapper[4675]: I0219 08:49:46.627345 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrwn" event={"ID":"50ddb4cd-7719-4f7a-a71f-a8bd832236e0","Type":"ContainerDied","Data":"662cb6923858d2161f5e5090d0b7444133ecb02fa7705c65598b4204c64a4f08"} Feb 19 08:49:46 crc kubenswrapper[4675]: I0219 08:49:46.652811 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gmkvv" podStartSLOduration=2.079559237 podStartE2EDuration="3.652791147s" podCreationTimestamp="2026-02-19 08:49:43 +0000 UTC" firstStartedPulling="2026-02-19 08:49:44.57545553 +0000 UTC m=+366.202545798" lastFinishedPulling="2026-02-19 08:49:46.14868744 +0000 UTC m=+367.775777708" observedRunningTime="2026-02-19 08:49:46.649287759 +0000 UTC m=+368.276378017" watchObservedRunningTime="2026-02-19 08:49:46.652791147 +0000 UTC m=+368.279881415" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.356248 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4r2wp"] Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.357521 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.377168 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4r2wp"] Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.486967 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6a437632-b571-4317-a2aa-f8540c878dab-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.487013 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6a437632-b571-4317-a2aa-f8540c878dab-registry-certificates\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.487043 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.487069 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-registry-tls\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.487220 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a437632-b571-4317-a2aa-f8540c878dab-trusted-ca\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.487304 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6a437632-b571-4317-a2aa-f8540c878dab-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.487342 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlpvz\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-kube-api-access-vlpvz\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.487415 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-bound-sa-token\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.560392 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.588384 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6a437632-b571-4317-a2aa-f8540c878dab-registry-certificates\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.588425 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6a437632-b571-4317-a2aa-f8540c878dab-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.588450 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-registry-tls\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.588492 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a437632-b571-4317-a2aa-f8540c878dab-trusted-ca\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.588529 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6a437632-b571-4317-a2aa-f8540c878dab-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.588554 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlpvz\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-kube-api-access-vlpvz\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.588589 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-bound-sa-token\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.589148 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6a437632-b571-4317-a2aa-f8540c878dab-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.589883 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6a437632-b571-4317-a2aa-f8540c878dab-registry-certificates\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.590081 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a437632-b571-4317-a2aa-f8540c878dab-trusted-ca\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.594888 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-registry-tls\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.595724 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6a437632-b571-4317-a2aa-f8540c878dab-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.607236 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-bound-sa-token\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.610246 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlpvz\" (UniqueName: \"kubernetes.io/projected/6a437632-b571-4317-a2aa-f8540c878dab-kube-api-access-vlpvz\") pod \"image-registry-66df7c8f76-4r2wp\" (UID: \"6a437632-b571-4317-a2aa-f8540c878dab\") " pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.633806 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrwn" event={"ID":"50ddb4cd-7719-4f7a-a71f-a8bd832236e0","Type":"ContainerStarted","Data":"3107474b2989d8ae5ea807b36633c5a71100194ed8cbb7a47ad7ea4ed35528bf"} Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.656661 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mvrwn" podStartSLOduration=2.07962902 podStartE2EDuration="4.65662578s" podCreationTimestamp="2026-02-19 08:49:43 +0000 UTC" firstStartedPulling="2026-02-19 08:49:44.583573918 +0000 UTC m=+366.210664206" lastFinishedPulling="2026-02-19 08:49:47.160570688 +0000 UTC m=+368.787660966" observedRunningTime="2026-02-19 08:49:47.654631325 +0000 UTC m=+369.281721593" watchObservedRunningTime="2026-02-19 08:49:47.65662578 +0000 UTC m=+369.283716038" Feb 19 08:49:47 crc kubenswrapper[4675]: I0219 08:49:47.674127 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:48 crc kubenswrapper[4675]: I0219 08:49:48.110031 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4r2wp"] Feb 19 08:49:48 crc kubenswrapper[4675]: I0219 08:49:48.640809 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" event={"ID":"6a437632-b571-4317-a2aa-f8540c878dab","Type":"ContainerStarted","Data":"a3cf2004929a4e9a5dfa2568ae245fae3776cac4365f755438a7ba1e0ee6a948"} Feb 19 08:49:48 crc kubenswrapper[4675]: I0219 08:49:48.641510 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" event={"ID":"6a437632-b571-4317-a2aa-f8540c878dab","Type":"ContainerStarted","Data":"8bc741efcf5775399b6c55a9c8139256d870df960b00527b9844327909bc3905"} Feb 19 08:49:48 crc kubenswrapper[4675]: I0219 08:49:48.641529 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:49:48 crc kubenswrapper[4675]: I0219 08:49:48.658266 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" podStartSLOduration=1.65824752 podStartE2EDuration="1.65824752s" podCreationTimestamp="2026-02-19 08:49:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:49:48.658110277 +0000 UTC m=+370.285200545" watchObservedRunningTime="2026-02-19 08:49:48.65824752 +0000 UTC m=+370.285337788" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.166738 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.167459 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.212438 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.329393 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.331114 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.388167 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.699211 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fg4v6" Feb 19 08:49:51 crc kubenswrapper[4675]: I0219 08:49:51.702761 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8xtr4" Feb 19 08:49:53 crc kubenswrapper[4675]: I0219 08:49:53.738305 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:53 crc kubenswrapper[4675]: I0219 08:49:53.738777 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:53 crc kubenswrapper[4675]: I0219 08:49:53.780006 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:49:53 crc kubenswrapper[4675]: I0219 08:49:53.931345 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:53 crc kubenswrapper[4675]: I0219 08:49:53.931417 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:53 crc kubenswrapper[4675]: I0219 08:49:53.968542 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:54 crc kubenswrapper[4675]: I0219 08:49:54.717600 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mvrwn" Feb 19 08:49:54 crc kubenswrapper[4675]: I0219 08:49:54.718009 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gmkvv" Feb 19 08:50:07 crc kubenswrapper[4675]: I0219 08:50:07.679383 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4r2wp" Feb 19 08:50:07 crc kubenswrapper[4675]: I0219 08:50:07.724295 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jfck7"] Feb 19 08:50:11 crc kubenswrapper[4675]: I0219 08:50:11.744332 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:50:11 crc kubenswrapper[4675]: I0219 08:50:11.745089 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:50:32 crc kubenswrapper[4675]: I0219 08:50:32.764234 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" podUID="08306565-7dde-46be-b2de-d813dfdd976a" containerName="registry" containerID="cri-o://20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108" gracePeriod=30 Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.135834 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.261380 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-registry-tls\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.261846 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-registry-certificates\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.261933 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-trusted-ca\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.261980 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/08306565-7dde-46be-b2de-d813dfdd976a-ca-trust-extracted\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.262318 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.262398 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mzqf\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-kube-api-access-8mzqf\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.262472 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/08306565-7dde-46be-b2de-d813dfdd976a-installation-pull-secrets\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.262570 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-bound-sa-token\") pod \"08306565-7dde-46be-b2de-d813dfdd976a\" (UID: \"08306565-7dde-46be-b2de-d813dfdd976a\") " Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.262960 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.262966 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.263235 4675 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.263254 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08306565-7dde-46be-b2de-d813dfdd976a-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.270134 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-kube-api-access-8mzqf" (OuterVolumeSpecName: "kube-api-access-8mzqf") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "kube-api-access-8mzqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.274286 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.274808 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08306565-7dde-46be-b2de-d813dfdd976a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.274983 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.275069 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.295384 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08306565-7dde-46be-b2de-d813dfdd976a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "08306565-7dde-46be-b2de-d813dfdd976a" (UID: "08306565-7dde-46be-b2de-d813dfdd976a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.363926 4675 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.363953 4675 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/08306565-7dde-46be-b2de-d813dfdd976a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.363965 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mzqf\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-kube-api-access-8mzqf\") on node \"crc\" DevicePath \"\"" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.363975 4675 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/08306565-7dde-46be-b2de-d813dfdd976a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.363984 4675 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/08306565-7dde-46be-b2de-d813dfdd976a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.885457 4675 generic.go:334] "Generic (PLEG): container finished" podID="08306565-7dde-46be-b2de-d813dfdd976a" containerID="20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108" exitCode=0 Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.885575 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" event={"ID":"08306565-7dde-46be-b2de-d813dfdd976a","Type":"ContainerDied","Data":"20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108"} Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.885660 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" event={"ID":"08306565-7dde-46be-b2de-d813dfdd976a","Type":"ContainerDied","Data":"98cd179ce5898a8d4c3bb46a00f1110934a579e96fccd4c949f1c04282a3a007"} Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.885686 4675 scope.go:117] "RemoveContainer" containerID="20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.887735 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jfck7" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.909506 4675 scope.go:117] "RemoveContainer" containerID="20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108" Feb 19 08:50:33 crc kubenswrapper[4675]: E0219 08:50:33.910034 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108\": container with ID starting with 20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108 not found: ID does not exist" containerID="20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.910080 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108"} err="failed to get container status \"20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108\": rpc error: code = NotFound desc = could not find container \"20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108\": container with ID starting with 20b705aba858c10273028b2ec0b3a878cd4f61a2e687244f9969a149e3542108 not found: ID does not exist" Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.927111 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jfck7"] Feb 19 08:50:33 crc kubenswrapper[4675]: I0219 08:50:33.934111 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jfck7"] Feb 19 08:50:35 crc kubenswrapper[4675]: I0219 08:50:35.110438 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08306565-7dde-46be-b2de-d813dfdd976a" path="/var/lib/kubelet/pods/08306565-7dde-46be-b2de-d813dfdd976a/volumes" Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.743904 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.744356 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.744437 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.745287 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ff7142b6a17f3a1cc3ed6d7db969da62d4d6cf49631898d5a6c5d25fdadb0c34"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.745364 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://ff7142b6a17f3a1cc3ed6d7db969da62d4d6cf49631898d5a6c5d25fdadb0c34" gracePeriod=600 Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.936197 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="ff7142b6a17f3a1cc3ed6d7db969da62d4d6cf49631898d5a6c5d25fdadb0c34" exitCode=0 Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.936401 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"ff7142b6a17f3a1cc3ed6d7db969da62d4d6cf49631898d5a6c5d25fdadb0c34"} Feb 19 08:50:41 crc kubenswrapper[4675]: I0219 08:50:41.936886 4675 scope.go:117] "RemoveContainer" containerID="af0d5043899af424124c2db0f92d0da93815d327358480b8fa5aeb372733f33e" Feb 19 08:50:42 crc kubenswrapper[4675]: I0219 08:50:42.946002 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"011bd75815a488feaac53d8c8b5eb25ad12052f61153f59a628707c37551dfbb"} Feb 19 08:53:11 crc kubenswrapper[4675]: I0219 08:53:11.743606 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:53:11 crc kubenswrapper[4675]: I0219 08:53:11.744338 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:53:41 crc kubenswrapper[4675]: I0219 08:53:41.743788 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:53:41 crc kubenswrapper[4675]: I0219 08:53:41.745139 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:54:11 crc kubenswrapper[4675]: I0219 08:54:11.744245 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:54:11 crc kubenswrapper[4675]: I0219 08:54:11.745097 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:54:11 crc kubenswrapper[4675]: I0219 08:54:11.745158 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:54:11 crc kubenswrapper[4675]: I0219 08:54:11.745795 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"011bd75815a488feaac53d8c8b5eb25ad12052f61153f59a628707c37551dfbb"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 08:54:11 crc kubenswrapper[4675]: I0219 08:54:11.745917 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://011bd75815a488feaac53d8c8b5eb25ad12052f61153f59a628707c37551dfbb" gracePeriod=600 Feb 19 08:54:12 crc kubenswrapper[4675]: I0219 08:54:12.118352 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="011bd75815a488feaac53d8c8b5eb25ad12052f61153f59a628707c37551dfbb" exitCode=0 Feb 19 08:54:12 crc kubenswrapper[4675]: I0219 08:54:12.118397 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"011bd75815a488feaac53d8c8b5eb25ad12052f61153f59a628707c37551dfbb"} Feb 19 08:54:12 crc kubenswrapper[4675]: I0219 08:54:12.118429 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"8ee4c969741bfa1ae0382fcd0dac3640f8d7616539fc52b4655b88e929fd6acf"} Feb 19 08:54:12 crc kubenswrapper[4675]: I0219 08:54:12.118445 4675 scope.go:117] "RemoveContainer" containerID="ff7142b6a17f3a1cc3ed6d7db969da62d4d6cf49631898d5a6c5d25fdadb0c34" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.947703 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-z9x24"] Feb 19 08:54:29 crc kubenswrapper[4675]: E0219 08:54:29.948604 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08306565-7dde-46be-b2de-d813dfdd976a" containerName="registry" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.948618 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="08306565-7dde-46be-b2de-d813dfdd976a" containerName="registry" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.948759 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="08306565-7dde-46be-b2de-d813dfdd976a" containerName="registry" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.949130 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.954332 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.954495 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.954499 4675 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tz9jq" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.957500 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-mfc4d"] Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.958220 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-mfc4d" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.959698 4675 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-csd55" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.963121 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-z9x24"] Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.974790 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-mfc4d"] Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.989304 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-5sxn8"] Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.990550 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.992686 4675 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-p9mpn" Feb 19 08:54:29 crc kubenswrapper[4675]: I0219 08:54:29.996517 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-5sxn8"] Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.053051 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jk79\" (UniqueName: \"kubernetes.io/projected/68aab37f-51c0-4132-844e-bc47b4856f16-kube-api-access-8jk79\") pod \"cert-manager-cainjector-cf98fcc89-z9x24\" (UID: \"68aab37f-51c0-4132-844e-bc47b4856f16\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.154575 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzmdv\" (UniqueName: \"kubernetes.io/projected/bf9ebddc-98fd-4515-8354-f1d7ae9fda4d-kube-api-access-jzmdv\") pod \"cert-manager-webhook-687f57d79b-5sxn8\" (UID: \"bf9ebddc-98fd-4515-8354-f1d7ae9fda4d\") " pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.154805 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l52xg\" (UniqueName: \"kubernetes.io/projected/2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e-kube-api-access-l52xg\") pod \"cert-manager-858654f9db-mfc4d\" (UID: \"2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e\") " pod="cert-manager/cert-manager-858654f9db-mfc4d" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.154893 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jk79\" (UniqueName: \"kubernetes.io/projected/68aab37f-51c0-4132-844e-bc47b4856f16-kube-api-access-8jk79\") pod \"cert-manager-cainjector-cf98fcc89-z9x24\" (UID: \"68aab37f-51c0-4132-844e-bc47b4856f16\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.177631 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jk79\" (UniqueName: \"kubernetes.io/projected/68aab37f-51c0-4132-844e-bc47b4856f16-kube-api-access-8jk79\") pod \"cert-manager-cainjector-cf98fcc89-z9x24\" (UID: \"68aab37f-51c0-4132-844e-bc47b4856f16\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.255564 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l52xg\" (UniqueName: \"kubernetes.io/projected/2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e-kube-api-access-l52xg\") pod \"cert-manager-858654f9db-mfc4d\" (UID: \"2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e\") " pod="cert-manager/cert-manager-858654f9db-mfc4d" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.255632 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzmdv\" (UniqueName: \"kubernetes.io/projected/bf9ebddc-98fd-4515-8354-f1d7ae9fda4d-kube-api-access-jzmdv\") pod \"cert-manager-webhook-687f57d79b-5sxn8\" (UID: \"bf9ebddc-98fd-4515-8354-f1d7ae9fda4d\") " pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.270872 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.272155 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l52xg\" (UniqueName: \"kubernetes.io/projected/2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e-kube-api-access-l52xg\") pod \"cert-manager-858654f9db-mfc4d\" (UID: \"2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e\") " pod="cert-manager/cert-manager-858654f9db-mfc4d" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.273899 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzmdv\" (UniqueName: \"kubernetes.io/projected/bf9ebddc-98fd-4515-8354-f1d7ae9fda4d-kube-api-access-jzmdv\") pod \"cert-manager-webhook-687f57d79b-5sxn8\" (UID: \"bf9ebddc-98fd-4515-8354-f1d7ae9fda4d\") " pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.285454 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-mfc4d" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.306651 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.673773 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-z9x24"] Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.687185 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.727551 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-mfc4d"] Feb 19 08:54:30 crc kubenswrapper[4675]: W0219 08:54:30.730955 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2138c669_3d8b_45c6_9c6d_dfc5d9ef0d2e.slice/crio-0418ba0a4876ce82f0d8f47fe99d9161b52b7d0bb4ce291a2780f269ff9ec089 WatchSource:0}: Error finding container 0418ba0a4876ce82f0d8f47fe99d9161b52b7d0bb4ce291a2780f269ff9ec089: Status 404 returned error can't find the container with id 0418ba0a4876ce82f0d8f47fe99d9161b52b7d0bb4ce291a2780f269ff9ec089 Feb 19 08:54:30 crc kubenswrapper[4675]: W0219 08:54:30.731706 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf9ebddc_98fd_4515_8354_f1d7ae9fda4d.slice/crio-0b194544ece25e4121a3956007221348b914531091f829232c5c20b4631cc219 WatchSource:0}: Error finding container 0b194544ece25e4121a3956007221348b914531091f829232c5c20b4631cc219: Status 404 returned error can't find the container with id 0b194544ece25e4121a3956007221348b914531091f829232c5c20b4631cc219 Feb 19 08:54:30 crc kubenswrapper[4675]: I0219 08:54:30.732548 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-5sxn8"] Feb 19 08:54:31 crc kubenswrapper[4675]: I0219 08:54:31.216820 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-mfc4d" event={"ID":"2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e","Type":"ContainerStarted","Data":"0418ba0a4876ce82f0d8f47fe99d9161b52b7d0bb4ce291a2780f269ff9ec089"} Feb 19 08:54:31 crc kubenswrapper[4675]: I0219 08:54:31.218134 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" event={"ID":"bf9ebddc-98fd-4515-8354-f1d7ae9fda4d","Type":"ContainerStarted","Data":"0b194544ece25e4121a3956007221348b914531091f829232c5c20b4631cc219"} Feb 19 08:54:31 crc kubenswrapper[4675]: I0219 08:54:31.220247 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" event={"ID":"68aab37f-51c0-4132-844e-bc47b4856f16","Type":"ContainerStarted","Data":"4b60829d20a8775fc2c16f67dae532fe3698e5dca6047e4a1e85fa59d6e0a571"} Feb 19 08:54:35 crc kubenswrapper[4675]: I0219 08:54:35.240397 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-mfc4d" event={"ID":"2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e","Type":"ContainerStarted","Data":"0a6b57e5c4316322b82c3f9e4b0f0a965181949b8f16c810313a1bab85e34a92"} Feb 19 08:54:35 crc kubenswrapper[4675]: I0219 08:54:35.241849 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" event={"ID":"bf9ebddc-98fd-4515-8354-f1d7ae9fda4d","Type":"ContainerStarted","Data":"abb7dc6c56e0d525bef5f20ea3de6c91beaa15887e7734c546af7a25915f6264"} Feb 19 08:54:35 crc kubenswrapper[4675]: I0219 08:54:35.241891 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" Feb 19 08:54:35 crc kubenswrapper[4675]: I0219 08:54:35.243327 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" event={"ID":"68aab37f-51c0-4132-844e-bc47b4856f16","Type":"ContainerStarted","Data":"25576889f7171f5f05544fff5e643d08887bae499b4a5aba2fce3cd0b4146170"} Feb 19 08:54:35 crc kubenswrapper[4675]: I0219 08:54:35.255183 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-mfc4d" podStartSLOduration=2.823067184 podStartE2EDuration="6.2551657s" podCreationTimestamp="2026-02-19 08:54:29 +0000 UTC" firstStartedPulling="2026-02-19 08:54:30.732652325 +0000 UTC m=+652.359742593" lastFinishedPulling="2026-02-19 08:54:34.164750841 +0000 UTC m=+655.791841109" observedRunningTime="2026-02-19 08:54:35.253245586 +0000 UTC m=+656.880335874" watchObservedRunningTime="2026-02-19 08:54:35.2551657 +0000 UTC m=+656.882255968" Feb 19 08:54:35 crc kubenswrapper[4675]: I0219 08:54:35.267070 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-z9x24" podStartSLOduration=2.842905002 podStartE2EDuration="6.267047454s" podCreationTimestamp="2026-02-19 08:54:29 +0000 UTC" firstStartedPulling="2026-02-19 08:54:30.68699068 +0000 UTC m=+652.314080948" lastFinishedPulling="2026-02-19 08:54:34.111133132 +0000 UTC m=+655.738223400" observedRunningTime="2026-02-19 08:54:35.265706316 +0000 UTC m=+656.892796594" watchObservedRunningTime="2026-02-19 08:54:35.267047454 +0000 UTC m=+656.894137722" Feb 19 08:54:35 crc kubenswrapper[4675]: I0219 08:54:35.302671 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" podStartSLOduration=2.924670994 podStartE2EDuration="6.302626276s" podCreationTimestamp="2026-02-19 08:54:29 +0000 UTC" firstStartedPulling="2026-02-19 08:54:30.733618532 +0000 UTC m=+652.360708800" lastFinishedPulling="2026-02-19 08:54:34.111573814 +0000 UTC m=+655.738664082" observedRunningTime="2026-02-19 08:54:35.302471102 +0000 UTC m=+656.929561370" watchObservedRunningTime="2026-02-19 08:54:35.302626276 +0000 UTC m=+656.929716544" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.029046 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lkvd2"] Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.030076 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-controller" containerID="cri-o://5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.030497 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="sbdb" containerID="cri-o://8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.030543 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="nbdb" containerID="cri-o://d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.030581 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="northd" containerID="cri-o://4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.030625 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.030683 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-node" containerID="cri-o://472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.030732 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-acl-logging" containerID="cri-o://09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.067519 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" containerID="cri-o://0a39b552dc1baa97f2a5ddf350fead6cbe122845f858e49017a916ed0ee17b36" gracePeriod=30 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.271042 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/2.log" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.271662 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/1.log" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.271716 4675 generic.go:334] "Generic (PLEG): container finished" podID="4cc07c5a-23e4-478f-88a7-73d41e17a320" containerID="664fc053ea88093ccbbf46c3c42c961a4cfe2c8ce85fca2abe20390388336078" exitCode=2 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.271792 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerDied","Data":"664fc053ea88093ccbbf46c3c42c961a4cfe2c8ce85fca2abe20390388336078"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.271843 4675 scope.go:117] "RemoveContainer" containerID="413d7d143b32c07366a26d9b9280a9a549e8e12c614b8d75867c603f7b0c2064" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.272376 4675 scope.go:117] "RemoveContainer" containerID="664fc053ea88093ccbbf46c3c42c961a4cfe2c8ce85fca2abe20390388336078" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.272587 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-qmz2s_openshift-multus(4cc07c5a-23e4-478f-88a7-73d41e17a320)\"" pod="openshift-multus/multus-qmz2s" podUID="4cc07c5a-23e4-478f-88a7-73d41e17a320" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.274976 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovnkube-controller/3.log" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.281414 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovn-acl-logging/0.log" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.281858 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovn-controller/0.log" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282151 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="0a39b552dc1baa97f2a5ddf350fead6cbe122845f858e49017a916ed0ee17b36" exitCode=0 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282179 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0" exitCode=0 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282191 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e" exitCode=0 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282203 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a" exitCode=0 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282211 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb" exitCode=0 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282220 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830" exitCode=0 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282236 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d" exitCode=143 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282244 4675 generic.go:334] "Generic (PLEG): container finished" podID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerID="5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a" exitCode=143 Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282266 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"0a39b552dc1baa97f2a5ddf350fead6cbe122845f858e49017a916ed0ee17b36"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282295 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282306 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282319 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282331 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282343 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282356 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.282367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a"} Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.309618 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-5sxn8" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.318939 4675 scope.go:117] "RemoveContainer" containerID="1df0c0947b7f3a7d9b8a7f05a234cfccb8ba513585dbb7865f9fb1c531b8f32d" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.322386 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovn-acl-logging/0.log" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.322948 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovn-controller/0.log" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.323394 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.371748 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c9cpv"] Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.371944 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="sbdb" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.371956 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="sbdb" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.371967 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.371973 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.371983 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-ovn-metrics" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.371990 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-ovn-metrics" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.371998 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372004 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372013 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-acl-logging" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372020 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-acl-logging" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372028 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="nbdb" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372034 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="nbdb" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372044 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372051 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372061 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="northd" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372066 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="northd" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372076 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372090 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372098 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-node" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372105 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-node" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372114 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kubecfg-setup" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372120 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kubecfg-setup" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372208 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372218 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="nbdb" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372227 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372235 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-ovn-metrics" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372242 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372251 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-acl-logging" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372258 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="kube-rbac-proxy-node" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372264 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovn-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372270 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="northd" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372278 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="sbdb" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372364 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372375 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: E0219 08:54:40.372385 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372391 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372634 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.372660 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" containerName="ovnkube-controller" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.374050 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402659 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-ovn-kubernetes\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402741 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5ndr\" (UniqueName: \"kubernetes.io/projected/d95acb1e-cd60-48dd-af45-e190cd07723a-kube-api-access-z5ndr\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402772 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-netns\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402794 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-var-lib-openvswitch\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402845 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-log-socket\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402867 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-openvswitch\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402918 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-systemd\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402947 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-config\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402966 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-kubelet\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.402988 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-node-log\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403009 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-ovn\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403033 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-env-overrides\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403054 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-bin\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403074 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403103 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-systemd-units\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403169 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-etc-openvswitch\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403191 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-script-lib\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403208 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d95acb1e-cd60-48dd-af45-e190cd07723a-ovn-node-metrics-cert\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403224 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-slash\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403247 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-netd\") pod \"d95acb1e-cd60-48dd-af45-e190cd07723a\" (UID: \"d95acb1e-cd60-48dd-af45-e190cd07723a\") " Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403318 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-log-socket\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403339 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-etc-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403357 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403375 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m79jx\" (UniqueName: \"kubernetes.io/projected/58d17593-518d-455d-a17b-910a3f1ba7ef-kube-api-access-m79jx\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403395 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58d17593-518d-455d-a17b-910a3f1ba7ef-ovn-node-metrics-cert\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403412 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-ovn\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403430 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-systemd\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403445 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-ovnkube-script-lib\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403478 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-slash\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403492 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-node-log\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403506 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-run-netns\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403523 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-cni-bin\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403544 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403560 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403575 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-kubelet\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403589 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-systemd-units\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403608 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-var-lib-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403622 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-cni-netd\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403660 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-ovnkube-config\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403681 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-env-overrides\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.403777 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.404168 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.404228 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.404265 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-log-socket" (OuterVolumeSpecName: "log-socket") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.404284 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.404916 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.404943 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405053 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405080 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-slash" (OuterVolumeSpecName: "host-slash") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405118 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405142 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405170 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405197 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-node-log" (OuterVolumeSpecName: "node-log") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405233 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405428 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405540 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.405564 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.409240 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95acb1e-cd60-48dd-af45-e190cd07723a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.409282 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95acb1e-cd60-48dd-af45-e190cd07723a-kube-api-access-z5ndr" (OuterVolumeSpecName: "kube-api-access-z5ndr") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "kube-api-access-z5ndr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.419272 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d95acb1e-cd60-48dd-af45-e190cd07723a" (UID: "d95acb1e-cd60-48dd-af45-e190cd07723a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504538 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504592 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504613 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-kubelet\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504650 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-systemd-units\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504673 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-var-lib-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504693 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-cni-netd\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504712 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-ovnkube-config\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504726 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-env-overrides\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504717 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504747 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-log-socket\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504750 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-systemd-units\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504717 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504765 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-etc-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504750 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-cni-netd\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504792 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-var-lib-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504825 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-kubelet\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504854 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-log-socket\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504884 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-etc-openvswitch\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504953 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.504997 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m79jx\" (UniqueName: \"kubernetes.io/projected/58d17593-518d-455d-a17b-910a3f1ba7ef-kube-api-access-m79jx\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505029 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58d17593-518d-455d-a17b-910a3f1ba7ef-ovn-node-metrics-cert\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505057 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-ovn\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505079 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-run-ovn-kubernetes\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505111 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-systemd\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505115 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-ovn\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505086 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-run-systemd\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505166 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-ovnkube-script-lib\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505201 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-slash\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505222 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-node-log\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505244 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-run-netns\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505268 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-cni-bin\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505318 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-slash\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505363 4675 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-log-socket\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505378 4675 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505385 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-cni-bin\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505390 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-host-run-netns\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505464 4675 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505481 4675 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505469 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/58d17593-518d-455d-a17b-910a3f1ba7ef-node-log\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505493 4675 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505581 4675 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-node-log\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505600 4675 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505614 4675 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505629 4675 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505699 4675 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505714 4675 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505728 4675 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505740 4675 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d95acb1e-cd60-48dd-af45-e190cd07723a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505753 4675 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d95acb1e-cd60-48dd-af45-e190cd07723a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505766 4675 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-slash\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505778 4675 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505791 4675 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505804 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5ndr\" (UniqueName: \"kubernetes.io/projected/d95acb1e-cd60-48dd-af45-e190cd07723a-kube-api-access-z5ndr\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505816 4675 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505831 4675 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d95acb1e-cd60-48dd-af45-e190cd07723a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.505836 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-ovnkube-config\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.506039 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-ovnkube-script-lib\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.506347 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/58d17593-518d-455d-a17b-910a3f1ba7ef-env-overrides\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.508370 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/58d17593-518d-455d-a17b-910a3f1ba7ef-ovn-node-metrics-cert\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.518799 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m79jx\" (UniqueName: \"kubernetes.io/projected/58d17593-518d-455d-a17b-910a3f1ba7ef-kube-api-access-m79jx\") pod \"ovnkube-node-c9cpv\" (UID: \"58d17593-518d-455d-a17b-910a3f1ba7ef\") " pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: I0219 08:54:40.687346 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:40 crc kubenswrapper[4675]: W0219 08:54:40.703202 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58d17593_518d_455d_a17b_910a3f1ba7ef.slice/crio-e89fe347692dc4ad0c90841bce55167d3354bb234ba88b949ab234eda0477484 WatchSource:0}: Error finding container e89fe347692dc4ad0c90841bce55167d3354bb234ba88b949ab234eda0477484: Status 404 returned error can't find the container with id e89fe347692dc4ad0c90841bce55167d3354bb234ba88b949ab234eda0477484 Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.288045 4675 generic.go:334] "Generic (PLEG): container finished" podID="58d17593-518d-455d-a17b-910a3f1ba7ef" containerID="40b5a41ef0a64a602c47ee3544459abbce8b8d744e019135f0fa93f21a374ecb" exitCode=0 Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.288073 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerDied","Data":"40b5a41ef0a64a602c47ee3544459abbce8b8d744e019135f0fa93f21a374ecb"} Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.288127 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"e89fe347692dc4ad0c90841bce55167d3354bb234ba88b949ab234eda0477484"} Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.293569 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovn-acl-logging/0.log" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.294119 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lkvd2_d95acb1e-cd60-48dd-af45-e190cd07723a/ovn-controller/0.log" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.294742 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" event={"ID":"d95acb1e-cd60-48dd-af45-e190cd07723a","Type":"ContainerDied","Data":"805386fab984c5af04d7bf2953d3e47915f49059662b00a9cdde597229de5a30"} Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.294791 4675 scope.go:117] "RemoveContainer" containerID="0a39b552dc1baa97f2a5ddf350fead6cbe122845f858e49017a916ed0ee17b36" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.295011 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lkvd2" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.302767 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/2.log" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.316099 4675 scope.go:117] "RemoveContainer" containerID="8e6cb6432d751fedeacfa0043b825df4a8a76d9da0cb2234d6aed32bdbb846d0" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.394553 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lkvd2"] Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.397785 4675 scope.go:117] "RemoveContainer" containerID="d67721ee579d400a694c81bf99110ed53d9861c9ca29994a08fc22d08681df4e" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.404091 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lkvd2"] Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.425827 4675 scope.go:117] "RemoveContainer" containerID="4ed58cd3b0b784307b5e58d985d146145feca2536db11130840fae9075155e1a" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.445540 4675 scope.go:117] "RemoveContainer" containerID="9e882e1ab8d4d82a912ad24390e856ad56dbfce20485935ca6720d4e5199c0cb" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.459050 4675 scope.go:117] "RemoveContainer" containerID="472694d21dbb77e37d0409eab97c1e3f06061c5aca689e0906989f081f840830" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.475031 4675 scope.go:117] "RemoveContainer" containerID="09819e33d8a5116730bc1f91c00ab70457dbdce7e2dd0d129d9d57abeab1030d" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.490417 4675 scope.go:117] "RemoveContainer" containerID="5ce404db25e6ae3d33b2f75f1700646d0ea40d1580d5d9761effc791b0f3587a" Feb 19 08:54:41 crc kubenswrapper[4675]: I0219 08:54:41.522856 4675 scope.go:117] "RemoveContainer" containerID="85e414e55cfd7057da22b7caa0a79bbb5a2a071b582f3f946e1fe93a767fc7ae" Feb 19 08:54:42 crc kubenswrapper[4675]: I0219 08:54:42.310464 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"90bb3850e7858a9f88fe48e49d6052ad9c9350600d3fc35731ef089c380db301"} Feb 19 08:54:42 crc kubenswrapper[4675]: I0219 08:54:42.310770 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"29e1e660756acf4244bdb09156c1ccf7f1580f68c670ebc448140d58633fbdfc"} Feb 19 08:54:42 crc kubenswrapper[4675]: I0219 08:54:42.310783 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"7b3a5eb87ff2b75cba398f2e275a445c7ee82b63fa95adff8d62aa7966ba059a"} Feb 19 08:54:42 crc kubenswrapper[4675]: I0219 08:54:42.310792 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"22e1a9c516aa6d80f11dd5c25188b0c13f8f0554ddcf944b557abbb7fbd09531"} Feb 19 08:54:42 crc kubenswrapper[4675]: I0219 08:54:42.310801 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"ec565591c094daf0a487d676cd932f30b01f9d87866fab73f54188c52c593e61"} Feb 19 08:54:42 crc kubenswrapper[4675]: I0219 08:54:42.310810 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"1bac3a66791b220c01a4981276b544076281cd1954eb7925d2247132c2d85d5b"} Feb 19 08:54:43 crc kubenswrapper[4675]: I0219 08:54:43.110174 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d95acb1e-cd60-48dd-af45-e190cd07723a" path="/var/lib/kubelet/pods/d95acb1e-cd60-48dd-af45-e190cd07723a/volumes" Feb 19 08:54:44 crc kubenswrapper[4675]: I0219 08:54:44.324760 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"f99c29ea2598a428734df4c2ccfc2abd7aa27d6c180f1fcbe723329c5f81985a"} Feb 19 08:54:47 crc kubenswrapper[4675]: I0219 08:54:47.342599 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" event={"ID":"58d17593-518d-455d-a17b-910a3f1ba7ef","Type":"ContainerStarted","Data":"bdd29f69354d4737639026e6ae3776f9e8cfceed0a922aae1775036ed4b49769"} Feb 19 08:54:47 crc kubenswrapper[4675]: I0219 08:54:47.343134 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:47 crc kubenswrapper[4675]: I0219 08:54:47.343151 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:47 crc kubenswrapper[4675]: I0219 08:54:47.343161 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:47 crc kubenswrapper[4675]: I0219 08:54:47.371985 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:47 crc kubenswrapper[4675]: I0219 08:54:47.373807 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" podStartSLOduration=7.373785992 podStartE2EDuration="7.373785992s" podCreationTimestamp="2026-02-19 08:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:54:47.368926375 +0000 UTC m=+668.996016643" watchObservedRunningTime="2026-02-19 08:54:47.373785992 +0000 UTC m=+669.000876260" Feb 19 08:54:47 crc kubenswrapper[4675]: I0219 08:54:47.374843 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:54:52 crc kubenswrapper[4675]: I0219 08:54:52.103224 4675 scope.go:117] "RemoveContainer" containerID="664fc053ea88093ccbbf46c3c42c961a4cfe2c8ce85fca2abe20390388336078" Feb 19 08:54:52 crc kubenswrapper[4675]: E0219 08:54:52.103749 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-qmz2s_openshift-multus(4cc07c5a-23e4-478f-88a7-73d41e17a320)\"" pod="openshift-multus/multus-qmz2s" podUID="4cc07c5a-23e4-478f-88a7-73d41e17a320" Feb 19 08:55:06 crc kubenswrapper[4675]: I0219 08:55:06.103285 4675 scope.go:117] "RemoveContainer" containerID="664fc053ea88093ccbbf46c3c42c961a4cfe2c8ce85fca2abe20390388336078" Feb 19 08:55:06 crc kubenswrapper[4675]: I0219 08:55:06.446732 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qmz2s_4cc07c5a-23e4-478f-88a7-73d41e17a320/kube-multus/2.log" Feb 19 08:55:06 crc kubenswrapper[4675]: I0219 08:55:06.447034 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qmz2s" event={"ID":"4cc07c5a-23e4-478f-88a7-73d41e17a320","Type":"ContainerStarted","Data":"d8812119e0daa4fb931ff4b77a17b97fa8931b0e85b619c06a5b37a3450da087"} Feb 19 08:55:10 crc kubenswrapper[4675]: I0219 08:55:10.717240 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c9cpv" Feb 19 08:55:15 crc kubenswrapper[4675]: I0219 08:55:15.996577 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk"] Feb 19 08:55:15 crc kubenswrapper[4675]: I0219 08:55:15.998350 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.000247 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.009261 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk"] Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.102216 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd6hz\" (UniqueName: \"kubernetes.io/projected/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-kube-api-access-rd6hz\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.102286 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.102490 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.204174 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.204289 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd6hz\" (UniqueName: \"kubernetes.io/projected/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-kube-api-access-rd6hz\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.204321 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.205030 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.205163 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.224869 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd6hz\" (UniqueName: \"kubernetes.io/projected/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-kube-api-access-rd6hz\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.318500 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:16 crc kubenswrapper[4675]: I0219 08:55:16.508051 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk"] Feb 19 08:55:17 crc kubenswrapper[4675]: I0219 08:55:17.516497 4675 generic.go:334] "Generic (PLEG): container finished" podID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerID="83b4c7d56bf6a460fc72b86e07094279f11b5b70546f91a8be00c62495247029" exitCode=0 Feb 19 08:55:17 crc kubenswrapper[4675]: I0219 08:55:17.516599 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" event={"ID":"c07488ba-ffa0-479e-8c03-bb41ae1b24f3","Type":"ContainerDied","Data":"83b4c7d56bf6a460fc72b86e07094279f11b5b70546f91a8be00c62495247029"} Feb 19 08:55:17 crc kubenswrapper[4675]: I0219 08:55:17.519490 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" event={"ID":"c07488ba-ffa0-479e-8c03-bb41ae1b24f3","Type":"ContainerStarted","Data":"b959bf5b9cdcd7bea1df512a5d6345d24913a710301ee7a9f721308a4bb6de06"} Feb 19 08:55:19 crc kubenswrapper[4675]: I0219 08:55:19.531308 4675 generic.go:334] "Generic (PLEG): container finished" podID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerID="df1ca87c8611a7e8c732029080754448eea98594ec333d7ed592dfd7f4f815d8" exitCode=0 Feb 19 08:55:19 crc kubenswrapper[4675]: I0219 08:55:19.531601 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" event={"ID":"c07488ba-ffa0-479e-8c03-bb41ae1b24f3","Type":"ContainerDied","Data":"df1ca87c8611a7e8c732029080754448eea98594ec333d7ed592dfd7f4f815d8"} Feb 19 08:55:20 crc kubenswrapper[4675]: I0219 08:55:20.542079 4675 generic.go:334] "Generic (PLEG): container finished" podID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerID="b67b256f546d0aa5312d6351e7ac61944f697d5ef42eebecbd942e70493d56c9" exitCode=0 Feb 19 08:55:20 crc kubenswrapper[4675]: I0219 08:55:20.542168 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" event={"ID":"c07488ba-ffa0-479e-8c03-bb41ae1b24f3","Type":"ContainerDied","Data":"b67b256f546d0aa5312d6351e7ac61944f697d5ef42eebecbd942e70493d56c9"} Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.849486 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.883730 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-bundle\") pod \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.883789 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd6hz\" (UniqueName: \"kubernetes.io/projected/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-kube-api-access-rd6hz\") pod \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.883877 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-util\") pod \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\" (UID: \"c07488ba-ffa0-479e-8c03-bb41ae1b24f3\") " Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.885296 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-bundle" (OuterVolumeSpecName: "bundle") pod "c07488ba-ffa0-479e-8c03-bb41ae1b24f3" (UID: "c07488ba-ffa0-479e-8c03-bb41ae1b24f3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.890293 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-kube-api-access-rd6hz" (OuterVolumeSpecName: "kube-api-access-rd6hz") pod "c07488ba-ffa0-479e-8c03-bb41ae1b24f3" (UID: "c07488ba-ffa0-479e-8c03-bb41ae1b24f3"). InnerVolumeSpecName "kube-api-access-rd6hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.985447 4675 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:55:21 crc kubenswrapper[4675]: I0219 08:55:21.985479 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd6hz\" (UniqueName: \"kubernetes.io/projected/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-kube-api-access-rd6hz\") on node \"crc\" DevicePath \"\"" Feb 19 08:55:22 crc kubenswrapper[4675]: I0219 08:55:22.223312 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-util" (OuterVolumeSpecName: "util") pod "c07488ba-ffa0-479e-8c03-bb41ae1b24f3" (UID: "c07488ba-ffa0-479e-8c03-bb41ae1b24f3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:55:22 crc kubenswrapper[4675]: I0219 08:55:22.290894 4675 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c07488ba-ffa0-479e-8c03-bb41ae1b24f3-util\") on node \"crc\" DevicePath \"\"" Feb 19 08:55:22 crc kubenswrapper[4675]: I0219 08:55:22.558271 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" Feb 19 08:55:22 crc kubenswrapper[4675]: I0219 08:55:22.558263 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk" event={"ID":"c07488ba-ffa0-479e-8c03-bb41ae1b24f3","Type":"ContainerDied","Data":"b959bf5b9cdcd7bea1df512a5d6345d24913a710301ee7a9f721308a4bb6de06"} Feb 19 08:55:22 crc kubenswrapper[4675]: I0219 08:55:22.558440 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b959bf5b9cdcd7bea1df512a5d6345d24913a710301ee7a9f721308a4bb6de06" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.620360 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-mmtc8"] Feb 19 08:55:24 crc kubenswrapper[4675]: E0219 08:55:24.620936 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerName="pull" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.620949 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerName="pull" Feb 19 08:55:24 crc kubenswrapper[4675]: E0219 08:55:24.620965 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerName="extract" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.620971 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerName="extract" Feb 19 08:55:24 crc kubenswrapper[4675]: E0219 08:55:24.620984 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerName="util" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.620990 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerName="util" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.621078 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="c07488ba-ffa0-479e-8c03-bb41ae1b24f3" containerName="extract" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.621430 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.624214 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.624747 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.625715 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-qdw5h" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.633650 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-mmtc8"] Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.723374 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxrbt\" (UniqueName: \"kubernetes.io/projected/a6fb0b89-c8be-4b62-a92b-23197720e978-kube-api-access-jxrbt\") pod \"nmstate-operator-694c9596b7-mmtc8\" (UID: \"a6fb0b89-c8be-4b62-a92b-23197720e978\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.824169 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxrbt\" (UniqueName: \"kubernetes.io/projected/a6fb0b89-c8be-4b62-a92b-23197720e978-kube-api-access-jxrbt\") pod \"nmstate-operator-694c9596b7-mmtc8\" (UID: \"a6fb0b89-c8be-4b62-a92b-23197720e978\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.842110 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxrbt\" (UniqueName: \"kubernetes.io/projected/a6fb0b89-c8be-4b62-a92b-23197720e978-kube-api-access-jxrbt\") pod \"nmstate-operator-694c9596b7-mmtc8\" (UID: \"a6fb0b89-c8be-4b62-a92b-23197720e978\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" Feb 19 08:55:24 crc kubenswrapper[4675]: I0219 08:55:24.937257 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" Feb 19 08:55:25 crc kubenswrapper[4675]: I0219 08:55:25.343673 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-mmtc8"] Feb 19 08:55:25 crc kubenswrapper[4675]: I0219 08:55:25.575579 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" event={"ID":"a6fb0b89-c8be-4b62-a92b-23197720e978","Type":"ContainerStarted","Data":"7d142d4eed99b4b260d345a2aa205003a89ba587e31c926acdd3f7b2619a0ae0"} Feb 19 08:55:27 crc kubenswrapper[4675]: I0219 08:55:27.589620 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" event={"ID":"a6fb0b89-c8be-4b62-a92b-23197720e978","Type":"ContainerStarted","Data":"8c08cc7933b09e1fd46c04c01862e1baf1b9835ab6d5dffdaecb6a9d0519ec4e"} Feb 19 08:55:27 crc kubenswrapper[4675]: I0219 08:55:27.606955 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-mmtc8" podStartSLOduration=1.78075881 podStartE2EDuration="3.606933564s" podCreationTimestamp="2026-02-19 08:55:24 +0000 UTC" firstStartedPulling="2026-02-19 08:55:25.357251527 +0000 UTC m=+706.984341795" lastFinishedPulling="2026-02-19 08:55:27.183426281 +0000 UTC m=+708.810516549" observedRunningTime="2026-02-19 08:55:27.60218604 +0000 UTC m=+709.229276318" watchObservedRunningTime="2026-02-19 08:55:27.606933564 +0000 UTC m=+709.234023832" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.619470 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-92cqz"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.620863 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.622774 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-m2m4v" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.635232 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-92cqz"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.649914 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.650726 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.652509 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.676753 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.684555 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-k28cx"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.685693 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.696409 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8krc\" (UniqueName: \"kubernetes.io/projected/8a261a06-8401-488e-8b98-d885f1b694cf-kube-api-access-c8krc\") pod \"nmstate-metrics-58c85c668d-92cqz\" (UID: \"8a261a06-8401-488e-8b98-d885f1b694cf\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.766374 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.767189 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.777544 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.777744 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-gw78r" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.777861 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.782315 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.798115 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-nmstate-lock\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.798161 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-ovs-socket\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.798299 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3be88498-1577-460a-9431-4a4bc66eb217-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-chcgg\" (UID: \"3be88498-1577-460a-9431-4a4bc66eb217\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.798375 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8krc\" (UniqueName: \"kubernetes.io/projected/8a261a06-8401-488e-8b98-d885f1b694cf-kube-api-access-c8krc\") pod \"nmstate-metrics-58c85c668d-92cqz\" (UID: \"8a261a06-8401-488e-8b98-d885f1b694cf\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.798444 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqnpr\" (UniqueName: \"kubernetes.io/projected/d1523822-5fa6-446b-9cd1-3442536238f2-kube-api-access-vqnpr\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.798575 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h7gp\" (UniqueName: \"kubernetes.io/projected/3be88498-1577-460a-9431-4a4bc66eb217-kube-api-access-4h7gp\") pod \"nmstate-webhook-866bcb46dc-chcgg\" (UID: \"3be88498-1577-460a-9431-4a4bc66eb217\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.798663 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-dbus-socket\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.819475 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8krc\" (UniqueName: \"kubernetes.io/projected/8a261a06-8401-488e-8b98-d885f1b694cf-kube-api-access-c8krc\") pod \"nmstate-metrics-58c85c668d-92cqz\" (UID: \"8a261a06-8401-488e-8b98-d885f1b694cf\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900230 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-nmstate-lock\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900276 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6d09f36-de74-42bb-8b27-3ac3039ee87f-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900296 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e6d09f36-de74-42bb-8b27-3ac3039ee87f-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900322 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-ovs-socket\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900349 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3be88498-1577-460a-9431-4a4bc66eb217-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-chcgg\" (UID: \"3be88498-1577-460a-9431-4a4bc66eb217\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900378 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-nmstate-lock\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900436 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-ovs-socket\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900481 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqnpr\" (UniqueName: \"kubernetes.io/projected/d1523822-5fa6-446b-9cd1-3442536238f2-kube-api-access-vqnpr\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900573 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h7gp\" (UniqueName: \"kubernetes.io/projected/3be88498-1577-460a-9431-4a4bc66eb217-kube-api-access-4h7gp\") pod \"nmstate-webhook-866bcb46dc-chcgg\" (UID: \"3be88498-1577-460a-9431-4a4bc66eb217\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900623 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-dbus-socket\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900678 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44gdg\" (UniqueName: \"kubernetes.io/projected/e6d09f36-de74-42bb-8b27-3ac3039ee87f-kube-api-access-44gdg\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.900936 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d1523822-5fa6-446b-9cd1-3442536238f2-dbus-socket\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.910806 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3be88498-1577-460a-9431-4a4bc66eb217-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-chcgg\" (UID: \"3be88498-1577-460a-9431-4a4bc66eb217\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.919983 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqnpr\" (UniqueName: \"kubernetes.io/projected/d1523822-5fa6-446b-9cd1-3442536238f2-kube-api-access-vqnpr\") pod \"nmstate-handler-k28cx\" (UID: \"d1523822-5fa6-446b-9cd1-3442536238f2\") " pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.933378 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h7gp\" (UniqueName: \"kubernetes.io/projected/3be88498-1577-460a-9431-4a4bc66eb217-kube-api-access-4h7gp\") pod \"nmstate-webhook-866bcb46dc-chcgg\" (UID: \"3be88498-1577-460a-9431-4a4bc66eb217\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.940100 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.964898 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.983018 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6c5487495f-hdpxn"] Feb 19 08:55:28 crc kubenswrapper[4675]: I0219 08:55:28.985353 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.000498 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c5487495f-hdpxn"] Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.001236 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.009453 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44gdg\" (UniqueName: \"kubernetes.io/projected/e6d09f36-de74-42bb-8b27-3ac3039ee87f-kube-api-access-44gdg\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.009513 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6d09f36-de74-42bb-8b27-3ac3039ee87f-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.009543 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e6d09f36-de74-42bb-8b27-3ac3039ee87f-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: E0219 08:55:29.009801 4675 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 19 08:55:29 crc kubenswrapper[4675]: E0219 08:55:29.009874 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6d09f36-de74-42bb-8b27-3ac3039ee87f-plugin-serving-cert podName:e6d09f36-de74-42bb-8b27-3ac3039ee87f nodeName:}" failed. No retries permitted until 2026-02-19 08:55:29.509855791 +0000 UTC m=+711.136946059 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/e6d09f36-de74-42bb-8b27-3ac3039ee87f-plugin-serving-cert") pod "nmstate-console-plugin-5c78fc5d65-f84lw" (UID: "e6d09f36-de74-42bb-8b27-3ac3039ee87f") : secret "plugin-serving-cert" not found Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.010587 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e6d09f36-de74-42bb-8b27-3ac3039ee87f-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.030026 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44gdg\" (UniqueName: \"kubernetes.io/projected/e6d09f36-de74-42bb-8b27-3ac3039ee87f-kube-api-access-44gdg\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.112334 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-service-ca\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.112367 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-config\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.112398 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-oauth-serving-cert\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.112429 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-trusted-ca-bundle\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.112543 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ps79\" (UniqueName: \"kubernetes.io/projected/ebc9b9ba-38e1-4deb-9c48-4760916b8738-kube-api-access-4ps79\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.112700 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-serving-cert\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.112737 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-oauth-config\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.198881 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-92cqz"] Feb 19 08:55:29 crc kubenswrapper[4675]: W0219 08:55:29.203909 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a261a06_8401_488e_8b98_d885f1b694cf.slice/crio-a83009f517b6c848cca3088e253f7ef448088c8a60bfb376d22aada9bbe6cfaa WatchSource:0}: Error finding container a83009f517b6c848cca3088e253f7ef448088c8a60bfb376d22aada9bbe6cfaa: Status 404 returned error can't find the container with id a83009f517b6c848cca3088e253f7ef448088c8a60bfb376d22aada9bbe6cfaa Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.214533 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-service-ca\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.214569 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-config\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.214663 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-oauth-serving-cert\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.214723 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-trusted-ca-bundle\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.214745 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ps79\" (UniqueName: \"kubernetes.io/projected/ebc9b9ba-38e1-4deb-9c48-4760916b8738-kube-api-access-4ps79\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.214777 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-serving-cert\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.214799 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-oauth-config\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.216204 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-service-ca\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.216675 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-config\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.217161 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-oauth-serving-cert\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.218015 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebc9b9ba-38e1-4deb-9c48-4760916b8738-trusted-ca-bundle\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.218611 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-oauth-config\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.218820 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebc9b9ba-38e1-4deb-9c48-4760916b8738-console-serving-cert\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.231729 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ps79\" (UniqueName: \"kubernetes.io/projected/ebc9b9ba-38e1-4deb-9c48-4760916b8738-kube-api-access-4ps79\") pod \"console-6c5487495f-hdpxn\" (UID: \"ebc9b9ba-38e1-4deb-9c48-4760916b8738\") " pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.314619 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.367230 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg"] Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.497194 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c5487495f-hdpxn"] Feb 19 08:55:29 crc kubenswrapper[4675]: W0219 08:55:29.502348 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebc9b9ba_38e1_4deb_9c48_4760916b8738.slice/crio-9008a976f527556dca3a0d09c35fc5c78b1bdb0c8052a07902ab7259438cfd03 WatchSource:0}: Error finding container 9008a976f527556dca3a0d09c35fc5c78b1bdb0c8052a07902ab7259438cfd03: Status 404 returned error can't find the container with id 9008a976f527556dca3a0d09c35fc5c78b1bdb0c8052a07902ab7259438cfd03 Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.518164 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6d09f36-de74-42bb-8b27-3ac3039ee87f-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.521614 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6d09f36-de74-42bb-8b27-3ac3039ee87f-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-f84lw\" (UID: \"e6d09f36-de74-42bb-8b27-3ac3039ee87f\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.604557 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" event={"ID":"8a261a06-8401-488e-8b98-d885f1b694cf","Type":"ContainerStarted","Data":"a83009f517b6c848cca3088e253f7ef448088c8a60bfb376d22aada9bbe6cfaa"} Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.605924 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" event={"ID":"3be88498-1577-460a-9431-4a4bc66eb217","Type":"ContainerStarted","Data":"62d4321c5b378cf92857ef6e4525cf1bb5d05f5dca7aebfbd7479ab122e9c804"} Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.607772 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-k28cx" event={"ID":"d1523822-5fa6-446b-9cd1-3442536238f2","Type":"ContainerStarted","Data":"c041fc1351cef7ef3b584eafd697358dd89a17470c4d931817e2343fc897b046"} Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.608891 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c5487495f-hdpxn" event={"ID":"ebc9b9ba-38e1-4deb-9c48-4760916b8738","Type":"ContainerStarted","Data":"9008a976f527556dca3a0d09c35fc5c78b1bdb0c8052a07902ab7259438cfd03"} Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.693813 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" Feb 19 08:55:29 crc kubenswrapper[4675]: I0219 08:55:29.893724 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw"] Feb 19 08:55:30 crc kubenswrapper[4675]: I0219 08:55:30.616620 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c5487495f-hdpxn" event={"ID":"ebc9b9ba-38e1-4deb-9c48-4760916b8738","Type":"ContainerStarted","Data":"1035f8437aef1511dcd7ea506f68a21b8f9a40292f84b231e95852792113b8e1"} Feb 19 08:55:30 crc kubenswrapper[4675]: I0219 08:55:30.618596 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" event={"ID":"e6d09f36-de74-42bb-8b27-3ac3039ee87f","Type":"ContainerStarted","Data":"0f12da85596b557ed8cbccf804870a058523e787cc65c5a101726c623521a76d"} Feb 19 08:55:30 crc kubenswrapper[4675]: I0219 08:55:30.632000 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6c5487495f-hdpxn" podStartSLOduration=2.63197799 podStartE2EDuration="2.63197799s" podCreationTimestamp="2026-02-19 08:55:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:55:30.63021961 +0000 UTC m=+712.257309888" watchObservedRunningTime="2026-02-19 08:55:30.63197799 +0000 UTC m=+712.259068258" Feb 19 08:55:31 crc kubenswrapper[4675]: I0219 08:55:31.628017 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" event={"ID":"8a261a06-8401-488e-8b98-d885f1b694cf","Type":"ContainerStarted","Data":"f45adadfbc5fe9cbb513536f7266fdfcf3a9301380832f9feb3e665386d87276"} Feb 19 08:55:31 crc kubenswrapper[4675]: I0219 08:55:31.629385 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" event={"ID":"3be88498-1577-460a-9431-4a4bc66eb217","Type":"ContainerStarted","Data":"0568788629eb9b78b049202b4420b04a505a4061f488eee6ec40c8b42cb62994"} Feb 19 08:55:31 crc kubenswrapper[4675]: I0219 08:55:31.629661 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:55:31 crc kubenswrapper[4675]: I0219 08:55:31.632905 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-k28cx" event={"ID":"d1523822-5fa6-446b-9cd1-3442536238f2","Type":"ContainerStarted","Data":"28d0b9e045c7924d8443207cba0afbf3be46470f8f028ee1def509390f47cef9"} Feb 19 08:55:31 crc kubenswrapper[4675]: I0219 08:55:31.648722 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" podStartSLOduration=1.7283735789999999 podStartE2EDuration="3.648701113s" podCreationTimestamp="2026-02-19 08:55:28 +0000 UTC" firstStartedPulling="2026-02-19 08:55:29.375735272 +0000 UTC m=+711.002825540" lastFinishedPulling="2026-02-19 08:55:31.296062806 +0000 UTC m=+712.923153074" observedRunningTime="2026-02-19 08:55:31.64361743 +0000 UTC m=+713.270707708" watchObservedRunningTime="2026-02-19 08:55:31.648701113 +0000 UTC m=+713.275791381" Feb 19 08:55:31 crc kubenswrapper[4675]: I0219 08:55:31.672938 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-k28cx" podStartSLOduration=1.427349194 podStartE2EDuration="3.672912105s" podCreationTimestamp="2026-02-19 08:55:28 +0000 UTC" firstStartedPulling="2026-02-19 08:55:29.044764203 +0000 UTC m=+710.671854471" lastFinishedPulling="2026-02-19 08:55:31.290327114 +0000 UTC m=+712.917417382" observedRunningTime="2026-02-19 08:55:31.66637052 +0000 UTC m=+713.293460788" watchObservedRunningTime="2026-02-19 08:55:31.672912105 +0000 UTC m=+713.300002373" Feb 19 08:55:32 crc kubenswrapper[4675]: I0219 08:55:32.640558 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" event={"ID":"e6d09f36-de74-42bb-8b27-3ac3039ee87f","Type":"ContainerStarted","Data":"edd140155a62ef4cdd61994e3f4d022df76a140f117c71f09c9e19bb6b21e361"} Feb 19 08:55:32 crc kubenswrapper[4675]: I0219 08:55:32.641115 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:32 crc kubenswrapper[4675]: I0219 08:55:32.660269 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-f84lw" podStartSLOduration=2.153122308 podStartE2EDuration="4.660250942s" podCreationTimestamp="2026-02-19 08:55:28 +0000 UTC" firstStartedPulling="2026-02-19 08:55:29.900160746 +0000 UTC m=+711.527251014" lastFinishedPulling="2026-02-19 08:55:32.40728938 +0000 UTC m=+714.034379648" observedRunningTime="2026-02-19 08:55:32.659308955 +0000 UTC m=+714.286399213" watchObservedRunningTime="2026-02-19 08:55:32.660250942 +0000 UTC m=+714.287341210" Feb 19 08:55:34 crc kubenswrapper[4675]: I0219 08:55:34.654020 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" event={"ID":"8a261a06-8401-488e-8b98-d885f1b694cf","Type":"ContainerStarted","Data":"e39fb4bc82b4f069cb14312bf540d510cc21b459391b6d159123c16146d604c2"} Feb 19 08:55:34 crc kubenswrapper[4675]: I0219 08:55:34.670213 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-92cqz" podStartSLOduration=2.351561281 podStartE2EDuration="6.670196586s" podCreationTimestamp="2026-02-19 08:55:28 +0000 UTC" firstStartedPulling="2026-02-19 08:55:29.206677122 +0000 UTC m=+710.833767390" lastFinishedPulling="2026-02-19 08:55:33.525312427 +0000 UTC m=+715.152402695" observedRunningTime="2026-02-19 08:55:34.669427366 +0000 UTC m=+716.296517654" watchObservedRunningTime="2026-02-19 08:55:34.670196586 +0000 UTC m=+716.297286854" Feb 19 08:55:39 crc kubenswrapper[4675]: I0219 08:55:39.041011 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-k28cx" Feb 19 08:55:39 crc kubenswrapper[4675]: I0219 08:55:39.315264 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:39 crc kubenswrapper[4675]: I0219 08:55:39.315323 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:39 crc kubenswrapper[4675]: I0219 08:55:39.321805 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:39 crc kubenswrapper[4675]: I0219 08:55:39.687830 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6c5487495f-hdpxn" Feb 19 08:55:39 crc kubenswrapper[4675]: I0219 08:55:39.734598 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vg9l6"] Feb 19 08:55:48 crc kubenswrapper[4675]: I0219 08:55:48.975130 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-chcgg" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.614980 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8"] Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.616709 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.618749 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.625847 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8"] Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.780258 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.780392 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxxml\" (UniqueName: \"kubernetes.io/projected/7388a06e-7e03-4aec-8326-6bcfe1cc194a-kube-api-access-fxxml\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.780454 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.881892 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.881983 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.882018 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxxml\" (UniqueName: \"kubernetes.io/projected/7388a06e-7e03-4aec-8326-6bcfe1cc194a-kube-api-access-fxxml\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.882580 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.882633 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.901762 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxxml\" (UniqueName: \"kubernetes.io/projected/7388a06e-7e03-4aec-8326-6bcfe1cc194a-kube-api-access-fxxml\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:01 crc kubenswrapper[4675]: I0219 08:56:01.932318 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:02 crc kubenswrapper[4675]: I0219 08:56:02.154068 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8"] Feb 19 08:56:02 crc kubenswrapper[4675]: I0219 08:56:02.811705 4675 generic.go:334] "Generic (PLEG): container finished" podID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerID="98db963322f08ac9829f9daa7ccd4002b0dc0f8c3ebb07a96348f31f141fa5b6" exitCode=0 Feb 19 08:56:02 crc kubenswrapper[4675]: I0219 08:56:02.811864 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" event={"ID":"7388a06e-7e03-4aec-8326-6bcfe1cc194a","Type":"ContainerDied","Data":"98db963322f08ac9829f9daa7ccd4002b0dc0f8c3ebb07a96348f31f141fa5b6"} Feb 19 08:56:02 crc kubenswrapper[4675]: I0219 08:56:02.812203 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" event={"ID":"7388a06e-7e03-4aec-8326-6bcfe1cc194a","Type":"ContainerStarted","Data":"20e02639840f1b56871fc96f359c7e0c65bc11c76de9b9376ef8a68b40674835"} Feb 19 08:56:04 crc kubenswrapper[4675]: I0219 08:56:04.780096 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vg9l6" podUID="0b433b98-d626-4286-a89b-3618ba63b58a" containerName="console" containerID="cri-o://4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2" gracePeriod=15 Feb 19 08:56:04 crc kubenswrapper[4675]: I0219 08:56:04.832909 4675 generic.go:334] "Generic (PLEG): container finished" podID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerID="3bfc5baf7070f2c5250eecab148e5dcc5209187e7db06d49e79ca975dca3fe9d" exitCode=0 Feb 19 08:56:04 crc kubenswrapper[4675]: I0219 08:56:04.832956 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" event={"ID":"7388a06e-7e03-4aec-8326-6bcfe1cc194a","Type":"ContainerDied","Data":"3bfc5baf7070f2c5250eecab148e5dcc5209187e7db06d49e79ca975dca3fe9d"} Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.145402 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vg9l6_0b433b98-d626-4286-a89b-3618ba63b58a/console/0.log" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.145764 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.334269 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-oauth-serving-cert\") pod \"0b433b98-d626-4286-a89b-3618ba63b58a\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.334315 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-trusted-ca-bundle\") pod \"0b433b98-d626-4286-a89b-3618ba63b58a\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.334349 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-service-ca\") pod \"0b433b98-d626-4286-a89b-3618ba63b58a\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.334373 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-oauth-config\") pod \"0b433b98-d626-4286-a89b-3618ba63b58a\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.334415 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrzqz\" (UniqueName: \"kubernetes.io/projected/0b433b98-d626-4286-a89b-3618ba63b58a-kube-api-access-nrzqz\") pod \"0b433b98-d626-4286-a89b-3618ba63b58a\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.334495 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-console-config\") pod \"0b433b98-d626-4286-a89b-3618ba63b58a\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.334512 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-serving-cert\") pod \"0b433b98-d626-4286-a89b-3618ba63b58a\" (UID: \"0b433b98-d626-4286-a89b-3618ba63b58a\") " Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.335826 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0b433b98-d626-4286-a89b-3618ba63b58a" (UID: "0b433b98-d626-4286-a89b-3618ba63b58a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.335861 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-console-config" (OuterVolumeSpecName: "console-config") pod "0b433b98-d626-4286-a89b-3618ba63b58a" (UID: "0b433b98-d626-4286-a89b-3618ba63b58a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.336123 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b433b98-d626-4286-a89b-3618ba63b58a" (UID: "0b433b98-d626-4286-a89b-3618ba63b58a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.336519 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0b433b98-d626-4286-a89b-3618ba63b58a" (UID: "0b433b98-d626-4286-a89b-3618ba63b58a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.341142 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0b433b98-d626-4286-a89b-3618ba63b58a" (UID: "0b433b98-d626-4286-a89b-3618ba63b58a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.341520 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0b433b98-d626-4286-a89b-3618ba63b58a" (UID: "0b433b98-d626-4286-a89b-3618ba63b58a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.342439 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b433b98-d626-4286-a89b-3618ba63b58a-kube-api-access-nrzqz" (OuterVolumeSpecName: "kube-api-access-nrzqz") pod "0b433b98-d626-4286-a89b-3618ba63b58a" (UID: "0b433b98-d626-4286-a89b-3618ba63b58a"). InnerVolumeSpecName "kube-api-access-nrzqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.436423 4675 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-console-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.436795 4675 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.436806 4675 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.436817 4675 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.436826 4675 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b433b98-d626-4286-a89b-3618ba63b58a-service-ca\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.436835 4675 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b433b98-d626-4286-a89b-3618ba63b58a-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.436844 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrzqz\" (UniqueName: \"kubernetes.io/projected/0b433b98-d626-4286-a89b-3618ba63b58a-kube-api-access-nrzqz\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.846526 4675 generic.go:334] "Generic (PLEG): container finished" podID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerID="0b924c46d19c730e1bcad751994332915f1a021f2bf924b7ef50cf2c3bcf771b" exitCode=0 Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.846633 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" event={"ID":"7388a06e-7e03-4aec-8326-6bcfe1cc194a","Type":"ContainerDied","Data":"0b924c46d19c730e1bcad751994332915f1a021f2bf924b7ef50cf2c3bcf771b"} Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.850449 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vg9l6_0b433b98-d626-4286-a89b-3618ba63b58a/console/0.log" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.850506 4675 generic.go:334] "Generic (PLEG): container finished" podID="0b433b98-d626-4286-a89b-3618ba63b58a" containerID="4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2" exitCode=2 Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.850563 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vg9l6" event={"ID":"0b433b98-d626-4286-a89b-3618ba63b58a","Type":"ContainerDied","Data":"4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2"} Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.850594 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vg9l6" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.850610 4675 scope.go:117] "RemoveContainer" containerID="4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.850598 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vg9l6" event={"ID":"0b433b98-d626-4286-a89b-3618ba63b58a","Type":"ContainerDied","Data":"d89451d205cba4e1932026342cd8c433ec532c8d5a6745f1891cf489060e0a16"} Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.872250 4675 scope.go:117] "RemoveContainer" containerID="4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2" Feb 19 08:56:05 crc kubenswrapper[4675]: E0219 08:56:05.872722 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2\": container with ID starting with 4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2 not found: ID does not exist" containerID="4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.872771 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2"} err="failed to get container status \"4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2\": rpc error: code = NotFound desc = could not find container \"4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2\": container with ID starting with 4626195266e8f38eccf0dfd0ef3d7ab2a51eeeafa8b070e05652ac4d0fe28cf2 not found: ID does not exist" Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.882429 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vg9l6"] Feb 19 08:56:05 crc kubenswrapper[4675]: I0219 08:56:05.886473 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vg9l6"] Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.108923 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b433b98-d626-4286-a89b-3618ba63b58a" path="/var/lib/kubelet/pods/0b433b98-d626-4286-a89b-3618ba63b58a/volumes" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.127845 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.258426 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-util\") pod \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.258478 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxxml\" (UniqueName: \"kubernetes.io/projected/7388a06e-7e03-4aec-8326-6bcfe1cc194a-kube-api-access-fxxml\") pod \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.258517 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-bundle\") pod \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\" (UID: \"7388a06e-7e03-4aec-8326-6bcfe1cc194a\") " Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.259555 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-bundle" (OuterVolumeSpecName: "bundle") pod "7388a06e-7e03-4aec-8326-6bcfe1cc194a" (UID: "7388a06e-7e03-4aec-8326-6bcfe1cc194a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.262832 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7388a06e-7e03-4aec-8326-6bcfe1cc194a-kube-api-access-fxxml" (OuterVolumeSpecName: "kube-api-access-fxxml") pod "7388a06e-7e03-4aec-8326-6bcfe1cc194a" (UID: "7388a06e-7e03-4aec-8326-6bcfe1cc194a"). InnerVolumeSpecName "kube-api-access-fxxml". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.272125 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-util" (OuterVolumeSpecName: "util") pod "7388a06e-7e03-4aec-8326-6bcfe1cc194a" (UID: "7388a06e-7e03-4aec-8326-6bcfe1cc194a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.359843 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxxml\" (UniqueName: \"kubernetes.io/projected/7388a06e-7e03-4aec-8326-6bcfe1cc194a-kube-api-access-fxxml\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.360179 4675 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-util\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.360245 4675 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7388a06e-7e03-4aec-8326-6bcfe1cc194a-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.865552 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" event={"ID":"7388a06e-7e03-4aec-8326-6bcfe1cc194a","Type":"ContainerDied","Data":"20e02639840f1b56871fc96f359c7e0c65bc11c76de9b9376ef8a68b40674835"} Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.866001 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20e02639840f1b56871fc96f359c7e0c65bc11c76de9b9376ef8a68b40674835" Feb 19 08:56:07 crc kubenswrapper[4675]: I0219 08:56:07.865775 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8" Feb 19 08:56:12 crc kubenswrapper[4675]: I0219 08:56:12.874330 4675 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.180858 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9"] Feb 19 08:56:16 crc kubenswrapper[4675]: E0219 08:56:16.181417 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerName="extract" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.181429 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerName="extract" Feb 19 08:56:16 crc kubenswrapper[4675]: E0219 08:56:16.181444 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b433b98-d626-4286-a89b-3618ba63b58a" containerName="console" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.181451 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b433b98-d626-4286-a89b-3618ba63b58a" containerName="console" Feb 19 08:56:16 crc kubenswrapper[4675]: E0219 08:56:16.181459 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerName="util" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.181466 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerName="util" Feb 19 08:56:16 crc kubenswrapper[4675]: E0219 08:56:16.181473 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerName="pull" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.181479 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerName="pull" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.181582 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7388a06e-7e03-4aec-8326-6bcfe1cc194a" containerName="extract" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.181597 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b433b98-d626-4286-a89b-3618ba63b58a" containerName="console" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.181986 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.184537 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.184997 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.185254 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.185397 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.185565 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-vq8zb" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.270303 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e1476c5-559f-4da9-9092-ef6b08253b6a-apiservice-cert\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.270351 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnfgn\" (UniqueName: \"kubernetes.io/projected/8e1476c5-559f-4da9-9092-ef6b08253b6a-kube-api-access-dnfgn\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.270548 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e1476c5-559f-4da9-9092-ef6b08253b6a-webhook-cert\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.273844 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9"] Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.371682 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e1476c5-559f-4da9-9092-ef6b08253b6a-webhook-cert\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.371745 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e1476c5-559f-4da9-9092-ef6b08253b6a-apiservice-cert\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.371775 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnfgn\" (UniqueName: \"kubernetes.io/projected/8e1476c5-559f-4da9-9092-ef6b08253b6a-kube-api-access-dnfgn\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.379381 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e1476c5-559f-4da9-9092-ef6b08253b6a-apiservice-cert\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.392861 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e1476c5-559f-4da9-9092-ef6b08253b6a-webhook-cert\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.413408 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnfgn\" (UniqueName: \"kubernetes.io/projected/8e1476c5-559f-4da9-9092-ef6b08253b6a-kube-api-access-dnfgn\") pod \"metallb-operator-controller-manager-698bb84f65-v6wc9\" (UID: \"8e1476c5-559f-4da9-9092-ef6b08253b6a\") " pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.498265 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.499274 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw"] Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.500128 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.502542 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.502769 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.502954 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8qw74" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.569809 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw"] Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.676915 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7d966d2-c2fa-411d-8474-ab26f3cbea46-apiservice-cert\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.676979 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7d966d2-c2fa-411d-8474-ab26f3cbea46-webhook-cert\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.677054 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnl6n\" (UniqueName: \"kubernetes.io/projected/d7d966d2-c2fa-411d-8474-ab26f3cbea46-kube-api-access-cnl6n\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.782053 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnl6n\" (UniqueName: \"kubernetes.io/projected/d7d966d2-c2fa-411d-8474-ab26f3cbea46-kube-api-access-cnl6n\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.782483 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7d966d2-c2fa-411d-8474-ab26f3cbea46-apiservice-cert\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.782518 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7d966d2-c2fa-411d-8474-ab26f3cbea46-webhook-cert\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.787988 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7d966d2-c2fa-411d-8474-ab26f3cbea46-apiservice-cert\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.794316 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7d966d2-c2fa-411d-8474-ab26f3cbea46-webhook-cert\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.799688 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnl6n\" (UniqueName: \"kubernetes.io/projected/d7d966d2-c2fa-411d-8474-ab26f3cbea46-kube-api-access-cnl6n\") pod \"metallb-operator-webhook-server-85dcc8c87f-62fjw\" (UID: \"d7d966d2-c2fa-411d-8474-ab26f3cbea46\") " pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:16 crc kubenswrapper[4675]: I0219 08:56:16.858400 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:17 crc kubenswrapper[4675]: I0219 08:56:17.092179 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9"] Feb 19 08:56:17 crc kubenswrapper[4675]: W0219 08:56:17.103946 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e1476c5_559f_4da9_9092_ef6b08253b6a.slice/crio-ca7cabc5a90e5137623b86cc0536466aaaad9b6efa58ec00c50190d40cc91141 WatchSource:0}: Error finding container ca7cabc5a90e5137623b86cc0536466aaaad9b6efa58ec00c50190d40cc91141: Status 404 returned error can't find the container with id ca7cabc5a90e5137623b86cc0536466aaaad9b6efa58ec00c50190d40cc91141 Feb 19 08:56:17 crc kubenswrapper[4675]: I0219 08:56:17.192204 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw"] Feb 19 08:56:17 crc kubenswrapper[4675]: W0219 08:56:17.200858 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7d966d2_c2fa_411d_8474_ab26f3cbea46.slice/crio-9d595adace7924e035dd7c27070d9bfea89f2d70bcedac56b8635a98cd766395 WatchSource:0}: Error finding container 9d595adace7924e035dd7c27070d9bfea89f2d70bcedac56b8635a98cd766395: Status 404 returned error can't find the container with id 9d595adace7924e035dd7c27070d9bfea89f2d70bcedac56b8635a98cd766395 Feb 19 08:56:17 crc kubenswrapper[4675]: I0219 08:56:17.922473 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" event={"ID":"8e1476c5-559f-4da9-9092-ef6b08253b6a","Type":"ContainerStarted","Data":"ca7cabc5a90e5137623b86cc0536466aaaad9b6efa58ec00c50190d40cc91141"} Feb 19 08:56:17 crc kubenswrapper[4675]: I0219 08:56:17.923923 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" event={"ID":"d7d966d2-c2fa-411d-8474-ab26f3cbea46","Type":"ContainerStarted","Data":"9d595adace7924e035dd7c27070d9bfea89f2d70bcedac56b8635a98cd766395"} Feb 19 08:56:22 crc kubenswrapper[4675]: I0219 08:56:22.963470 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" event={"ID":"8e1476c5-559f-4da9-9092-ef6b08253b6a","Type":"ContainerStarted","Data":"0849c392f5a04899d1918c7cd72f4feb77adbb5c52b539b13188d7268b029f72"} Feb 19 08:56:22 crc kubenswrapper[4675]: I0219 08:56:22.964200 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:22 crc kubenswrapper[4675]: I0219 08:56:22.965069 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" event={"ID":"d7d966d2-c2fa-411d-8474-ab26f3cbea46","Type":"ContainerStarted","Data":"d6f4d071de9633c1b1af93641111e4ada5db22909f7a84248624e795f5e784b5"} Feb 19 08:56:22 crc kubenswrapper[4675]: I0219 08:56:22.965197 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:22 crc kubenswrapper[4675]: I0219 08:56:22.983457 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" podStartSLOduration=1.6163320410000002 podStartE2EDuration="6.983440145s" podCreationTimestamp="2026-02-19 08:56:16 +0000 UTC" firstStartedPulling="2026-02-19 08:56:17.107140015 +0000 UTC m=+758.734230273" lastFinishedPulling="2026-02-19 08:56:22.474248109 +0000 UTC m=+764.101338377" observedRunningTime="2026-02-19 08:56:22.980241739 +0000 UTC m=+764.607332007" watchObservedRunningTime="2026-02-19 08:56:22.983440145 +0000 UTC m=+764.610530413" Feb 19 08:56:36 crc kubenswrapper[4675]: I0219 08:56:36.863738 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" Feb 19 08:56:36 crc kubenswrapper[4675]: I0219 08:56:36.893442 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-85dcc8c87f-62fjw" podStartSLOduration=15.615310586 podStartE2EDuration="20.893424531s" podCreationTimestamp="2026-02-19 08:56:16 +0000 UTC" firstStartedPulling="2026-02-19 08:56:17.205179228 +0000 UTC m=+758.832269496" lastFinishedPulling="2026-02-19 08:56:22.483293173 +0000 UTC m=+764.110383441" observedRunningTime="2026-02-19 08:56:23.003582368 +0000 UTC m=+764.630672636" watchObservedRunningTime="2026-02-19 08:56:36.893424531 +0000 UTC m=+778.520514799" Feb 19 08:56:41 crc kubenswrapper[4675]: I0219 08:56:41.744158 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:56:41 crc kubenswrapper[4675]: I0219 08:56:41.744973 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:56:56 crc kubenswrapper[4675]: I0219 08:56:56.501285 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-698bb84f65-v6wc9" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.265266 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t"] Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.266853 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.269977 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jq47l" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.270008 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.277462 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-fj4cv"] Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.282098 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.287308 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t"] Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.288635 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.288947 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.348480 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bc8vc"] Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.349608 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.352874 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.352909 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-hh2nz" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.353090 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.353417 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355146 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-conf\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355194 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355227 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m86ll\" (UniqueName: \"kubernetes.io/projected/ee8232ef-f035-4ef4-b867-98508be2d0e7-kube-api-access-m86ll\") pod \"frr-k8s-webhook-server-78b44bf5bb-fjr4t\" (UID: \"ee8232ef-f035-4ef4-b867-98508be2d0e7\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355251 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-reloader\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355281 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtgwc\" (UniqueName: \"kubernetes.io/projected/e0bff342-a422-4fe7-b4cb-ca04e88bf905-kube-api-access-qtgwc\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355296 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-startup\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355312 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics-certs\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355351 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee8232ef-f035-4ef4-b867-98508be2d0e7-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-fjr4t\" (UID: \"ee8232ef-f035-4ef4-b867-98508be2d0e7\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.355370 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-sockets\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.363703 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-j4tln"] Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.364620 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.369240 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.383228 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-j4tln"] Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.456546 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-cert\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.456871 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtgwc\" (UniqueName: \"kubernetes.io/projected/e0bff342-a422-4fe7-b4cb-ca04e88bf905-kube-api-access-qtgwc\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.457277 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-startup\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458190 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics-certs\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458151 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-startup\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.458324 4675 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458353 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/75ab6240-327c-4ac4-94f2-3d4c44e1a664-metallb-excludel2\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.458508 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics-certs podName:e0bff342-a422-4fe7-b4cb-ca04e88bf905 nodeName:}" failed. No retries permitted until 2026-02-19 08:56:57.958477045 +0000 UTC m=+799.585567313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics-certs") pod "frr-k8s-fj4cv" (UID: "e0bff342-a422-4fe7-b4cb-ca04e88bf905") : secret "frr-k8s-certs-secret" not found Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458632 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458725 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tfxp\" (UniqueName: \"kubernetes.io/projected/47ca0e9d-940d-4952-9780-74e58570d98e-kube-api-access-9tfxp\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458806 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-metrics-certs\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458882 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee8232ef-f035-4ef4-b867-98508be2d0e7-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-fjr4t\" (UID: \"ee8232ef-f035-4ef4-b867-98508be2d0e7\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.458955 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-sockets\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.459025 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-conf\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.459101 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-metrics-certs\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.459175 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.459243 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flz5k\" (UniqueName: \"kubernetes.io/projected/75ab6240-327c-4ac4-94f2-3d4c44e1a664-kube-api-access-flz5k\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.459325 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m86ll\" (UniqueName: \"kubernetes.io/projected/ee8232ef-f035-4ef4-b867-98508be2d0e7-kube-api-access-m86ll\") pod \"frr-k8s-webhook-server-78b44bf5bb-fjr4t\" (UID: \"ee8232ef-f035-4ef4-b867-98508be2d0e7\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.459398 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-reloader\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.459704 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-reloader\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.459824 4675 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.459913 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee8232ef-f035-4ef4-b867-98508be2d0e7-cert podName:ee8232ef-f035-4ef4-b867-98508be2d0e7 nodeName:}" failed. No retries permitted until 2026-02-19 08:56:57.959899454 +0000 UTC m=+799.586989722 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ee8232ef-f035-4ef4-b867-98508be2d0e7-cert") pod "frr-k8s-webhook-server-78b44bf5bb-fjr4t" (UID: "ee8232ef-f035-4ef4-b867-98508be2d0e7") : secret "frr-k8s-webhook-server-cert" not found Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.460160 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-sockets\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.460396 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-frr-conf\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.460655 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.478044 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtgwc\" (UniqueName: \"kubernetes.io/projected/e0bff342-a422-4fe7-b4cb-ca04e88bf905-kube-api-access-qtgwc\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.478055 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m86ll\" (UniqueName: \"kubernetes.io/projected/ee8232ef-f035-4ef4-b867-98508be2d0e7-kube-api-access-m86ll\") pod \"frr-k8s-webhook-server-78b44bf5bb-fjr4t\" (UID: \"ee8232ef-f035-4ef4-b867-98508be2d0e7\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.561030 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-metrics-certs\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.561086 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flz5k\" (UniqueName: \"kubernetes.io/projected/75ab6240-327c-4ac4-94f2-3d4c44e1a664-kube-api-access-flz5k\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.561159 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-cert\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.561199 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/75ab6240-327c-4ac4-94f2-3d4c44e1a664-metallb-excludel2\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.561216 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.561236 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tfxp\" (UniqueName: \"kubernetes.io/projected/47ca0e9d-940d-4952-9780-74e58570d98e-kube-api-access-9tfxp\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.561259 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-metrics-certs\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.562155 4675 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.562205 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist podName:75ab6240-327c-4ac4-94f2-3d4c44e1a664 nodeName:}" failed. No retries permitted until 2026-02-19 08:56:58.062192367 +0000 UTC m=+799.689282635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist") pod "speaker-bc8vc" (UID: "75ab6240-327c-4ac4-94f2-3d4c44e1a664") : secret "metallb-memberlist" not found Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.562244 4675 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Feb 19 08:56:57 crc kubenswrapper[4675]: E0219 08:56:57.562332 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-metrics-certs podName:47ca0e9d-940d-4952-9780-74e58570d98e nodeName:}" failed. No retries permitted until 2026-02-19 08:56:58.06231555 +0000 UTC m=+799.689405818 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-metrics-certs") pod "controller-69bbfbf88f-j4tln" (UID: "47ca0e9d-940d-4952-9780-74e58570d98e") : secret "controller-certs-secret" not found Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.562823 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/75ab6240-327c-4ac4-94f2-3d4c44e1a664-metallb-excludel2\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.565980 4675 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.566164 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-metrics-certs\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.577095 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-cert\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.581020 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flz5k\" (UniqueName: \"kubernetes.io/projected/75ab6240-327c-4ac4-94f2-3d4c44e1a664-kube-api-access-flz5k\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.582221 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tfxp\" (UniqueName: \"kubernetes.io/projected/47ca0e9d-940d-4952-9780-74e58570d98e-kube-api-access-9tfxp\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.966728 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee8232ef-f035-4ef4-b867-98508be2d0e7-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-fjr4t\" (UID: \"ee8232ef-f035-4ef4-b867-98508be2d0e7\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.966870 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics-certs\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.971116 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e0bff342-a422-4fe7-b4cb-ca04e88bf905-metrics-certs\") pod \"frr-k8s-fj4cv\" (UID: \"e0bff342-a422-4fe7-b4cb-ca04e88bf905\") " pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:57 crc kubenswrapper[4675]: I0219 08:56:57.971487 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ee8232ef-f035-4ef4-b867-98508be2d0e7-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-fjr4t\" (UID: \"ee8232ef-f035-4ef4-b867-98508be2d0e7\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.067561 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-metrics-certs\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.067660 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:58 crc kubenswrapper[4675]: E0219 08:56:58.067834 4675 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 19 08:56:58 crc kubenswrapper[4675]: E0219 08:56:58.067887 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist podName:75ab6240-327c-4ac4-94f2-3d4c44e1a664 nodeName:}" failed. No retries permitted until 2026-02-19 08:56:59.067872833 +0000 UTC m=+800.694963101 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist") pod "speaker-bc8vc" (UID: "75ab6240-327c-4ac4-94f2-3d4c44e1a664") : secret "metallb-memberlist" not found Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.072411 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47ca0e9d-940d-4952-9780-74e58570d98e-metrics-certs\") pod \"controller-69bbfbf88f-j4tln\" (UID: \"47ca0e9d-940d-4952-9780-74e58570d98e\") " pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.186196 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.210609 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.325203 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.580420 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-j4tln"] Feb 19 08:56:58 crc kubenswrapper[4675]: I0219 08:56:58.665244 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t"] Feb 19 08:56:58 crc kubenswrapper[4675]: W0219 08:56:58.673258 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee8232ef_f035_4ef4_b867_98508be2d0e7.slice/crio-a25ba27e63cc336afa51043bfedf22e7d9a9fe1f0f16a9b6949ad62d646497a3 WatchSource:0}: Error finding container a25ba27e63cc336afa51043bfedf22e7d9a9fe1f0f16a9b6949ad62d646497a3: Status 404 returned error can't find the container with id a25ba27e63cc336afa51043bfedf22e7d9a9fe1f0f16a9b6949ad62d646497a3 Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.082439 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.089789 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/75ab6240-327c-4ac4-94f2-3d4c44e1a664-memberlist\") pod \"speaker-bc8vc\" (UID: \"75ab6240-327c-4ac4-94f2-3d4c44e1a664\") " pod="metallb-system/speaker-bc8vc" Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.158520 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" event={"ID":"ee8232ef-f035-4ef4-b867-98508be2d0e7","Type":"ContainerStarted","Data":"a25ba27e63cc336afa51043bfedf22e7d9a9fe1f0f16a9b6949ad62d646497a3"} Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.160050 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerStarted","Data":"ac555bf258ef57af214fa443f44fdfa53c822d2aafc6680112ce7869d2e5f4c5"} Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.162000 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-j4tln" event={"ID":"47ca0e9d-940d-4952-9780-74e58570d98e","Type":"ContainerStarted","Data":"ec042ffbd64a64b6a05fdf1681161006b4a5838467d3f7c2506cd751552cbd6a"} Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.162071 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-j4tln" event={"ID":"47ca0e9d-940d-4952-9780-74e58570d98e","Type":"ContainerStarted","Data":"6d96614061b3921c60882cc921294fad02b4fb703e480eb1ce5b7ef24f3d6f36"} Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.162087 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-j4tln" event={"ID":"47ca0e9d-940d-4952-9780-74e58570d98e","Type":"ContainerStarted","Data":"cfd9335d27e47dac6ce4f8ea2f9a27f067b4cdf5c0f06988e095f5724deebf0f"} Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.162153 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.165244 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bc8vc" Feb 19 08:56:59 crc kubenswrapper[4675]: I0219 08:56:59.181192 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-j4tln" podStartSLOduration=2.181163329 podStartE2EDuration="2.181163329s" podCreationTimestamp="2026-02-19 08:56:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:56:59.177599303 +0000 UTC m=+800.804689571" watchObservedRunningTime="2026-02-19 08:56:59.181163329 +0000 UTC m=+800.808253597" Feb 19 08:56:59 crc kubenswrapper[4675]: W0219 08:56:59.186738 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75ab6240_327c_4ac4_94f2_3d4c44e1a664.slice/crio-f67a3e3362adca1e43aac1c6d06ae9204e98507863d036a88202f509bee7f804 WatchSource:0}: Error finding container f67a3e3362adca1e43aac1c6d06ae9204e98507863d036a88202f509bee7f804: Status 404 returned error can't find the container with id f67a3e3362adca1e43aac1c6d06ae9204e98507863d036a88202f509bee7f804 Feb 19 08:57:00 crc kubenswrapper[4675]: I0219 08:57:00.179888 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bc8vc" event={"ID":"75ab6240-327c-4ac4-94f2-3d4c44e1a664","Type":"ContainerStarted","Data":"f76da6519cea9be61a6592e9bbd73e404ff6449e032ee59e4d3f8f120d964a36"} Feb 19 08:57:00 crc kubenswrapper[4675]: I0219 08:57:00.180332 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bc8vc" event={"ID":"75ab6240-327c-4ac4-94f2-3d4c44e1a664","Type":"ContainerStarted","Data":"5dd973e774418c1fc01d26cdbcda21f329b5b0dc959f20ad421949061e1c5311"} Feb 19 08:57:00 crc kubenswrapper[4675]: I0219 08:57:00.180351 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bc8vc" event={"ID":"75ab6240-327c-4ac4-94f2-3d4c44e1a664","Type":"ContainerStarted","Data":"f67a3e3362adca1e43aac1c6d06ae9204e98507863d036a88202f509bee7f804"} Feb 19 08:57:00 crc kubenswrapper[4675]: I0219 08:57:00.180545 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bc8vc" Feb 19 08:57:00 crc kubenswrapper[4675]: I0219 08:57:00.203246 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bc8vc" podStartSLOduration=3.203227083 podStartE2EDuration="3.203227083s" podCreationTimestamp="2026-02-19 08:56:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:57:00.198872685 +0000 UTC m=+801.825962963" watchObservedRunningTime="2026-02-19 08:57:00.203227083 +0000 UTC m=+801.830317351" Feb 19 08:57:07 crc kubenswrapper[4675]: I0219 08:57:07.261459 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" event={"ID":"ee8232ef-f035-4ef4-b867-98508be2d0e7","Type":"ContainerStarted","Data":"f93b3407403fb84856d2ed95eeda8baebdd3a42de59e563777631785162eee2a"} Feb 19 08:57:07 crc kubenswrapper[4675]: I0219 08:57:07.262138 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:57:07 crc kubenswrapper[4675]: I0219 08:57:07.263083 4675 generic.go:334] "Generic (PLEG): container finished" podID="e0bff342-a422-4fe7-b4cb-ca04e88bf905" containerID="d9e3372e77d3b8509651e9d72a31debe0b7d9928c91c8de01c9b9d7f2f90f22d" exitCode=0 Feb 19 08:57:07 crc kubenswrapper[4675]: I0219 08:57:07.263125 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerDied","Data":"d9e3372e77d3b8509651e9d72a31debe0b7d9928c91c8de01c9b9d7f2f90f22d"} Feb 19 08:57:07 crc kubenswrapper[4675]: I0219 08:57:07.276237 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" podStartSLOduration=2.623485988 podStartE2EDuration="10.276220315s" podCreationTimestamp="2026-02-19 08:56:57 +0000 UTC" firstStartedPulling="2026-02-19 08:56:58.67538195 +0000 UTC m=+800.302472218" lastFinishedPulling="2026-02-19 08:57:06.328116277 +0000 UTC m=+807.955206545" observedRunningTime="2026-02-19 08:57:07.274019826 +0000 UTC m=+808.901110094" watchObservedRunningTime="2026-02-19 08:57:07.276220315 +0000 UTC m=+808.903310583" Feb 19 08:57:08 crc kubenswrapper[4675]: I0219 08:57:08.280103 4675 generic.go:334] "Generic (PLEG): container finished" podID="e0bff342-a422-4fe7-b4cb-ca04e88bf905" containerID="8f253a11c048ba99103a0dde751a49181928a48978b03be7d4a4f95632a4cc76" exitCode=0 Feb 19 08:57:08 crc kubenswrapper[4675]: I0219 08:57:08.280249 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerDied","Data":"8f253a11c048ba99103a0dde751a49181928a48978b03be7d4a4f95632a4cc76"} Feb 19 08:57:08 crc kubenswrapper[4675]: I0219 08:57:08.329760 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-j4tln" Feb 19 08:57:09 crc kubenswrapper[4675]: I0219 08:57:09.169701 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bc8vc" Feb 19 08:57:09 crc kubenswrapper[4675]: I0219 08:57:09.289128 4675 generic.go:334] "Generic (PLEG): container finished" podID="e0bff342-a422-4fe7-b4cb-ca04e88bf905" containerID="d94e86260f70417210b97f98964b25e9aa05449e60dd9ce422e7ef899e1a2489" exitCode=0 Feb 19 08:57:09 crc kubenswrapper[4675]: I0219 08:57:09.289170 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerDied","Data":"d94e86260f70417210b97f98964b25e9aa05449e60dd9ce422e7ef899e1a2489"} Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.300506 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerStarted","Data":"1a0cd20ade0b1777f62436dd882670c9c4ae610b6d8ddd05d049b0cb18876a70"} Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.301023 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.301045 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerStarted","Data":"1166d6918635c14f18758730bf964c863abbc83c3b1ddc343f38c2293b2c1fc3"} Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.301063 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerStarted","Data":"f6d2c0162825015bcdbef4a6fdddb09391005a27ff8381e3c771c6b48ae75e84"} Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.301079 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerStarted","Data":"daf5837b15adcff2945a09a9be71567cb635dee81dab3c8e6db7b6ebd6f60860"} Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.301094 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerStarted","Data":"676f22d547d959ec77cc59a055163f2a157816425d8bbfca8a753ec33fcac611"} Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.301108 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fj4cv" event={"ID":"e0bff342-a422-4fe7-b4cb-ca04e88bf905","Type":"ContainerStarted","Data":"d08b0369a4f0d983525421e15cf3bfee50ccf70c0d56f7ee9f79efb42064b472"} Feb 19 08:57:10 crc kubenswrapper[4675]: I0219 08:57:10.323930 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-fj4cv" podStartSLOduration=5.372411388 podStartE2EDuration="13.323910833s" podCreationTimestamp="2026-02-19 08:56:57 +0000 UTC" firstStartedPulling="2026-02-19 08:56:58.394765384 +0000 UTC m=+800.021855652" lastFinishedPulling="2026-02-19 08:57:06.346264829 +0000 UTC m=+807.973355097" observedRunningTime="2026-02-19 08:57:10.321765055 +0000 UTC m=+811.948855333" watchObservedRunningTime="2026-02-19 08:57:10.323910833 +0000 UTC m=+811.951001101" Feb 19 08:57:11 crc kubenswrapper[4675]: I0219 08:57:11.743604 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:57:11 crc kubenswrapper[4675]: I0219 08:57:11.743702 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.122535 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vnslc"] Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.123352 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vnslc" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.125373 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-ht29t" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.126510 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.126558 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.144099 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vnslc"] Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.190795 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6grgs\" (UniqueName: \"kubernetes.io/projected/26967c39-74e8-4b89-89f8-e200c602de61-kube-api-access-6grgs\") pod \"openstack-operator-index-vnslc\" (UID: \"26967c39-74e8-4b89-89f8-e200c602de61\") " pod="openstack-operators/openstack-operator-index-vnslc" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.291718 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6grgs\" (UniqueName: \"kubernetes.io/projected/26967c39-74e8-4b89-89f8-e200c602de61-kube-api-access-6grgs\") pod \"openstack-operator-index-vnslc\" (UID: \"26967c39-74e8-4b89-89f8-e200c602de61\") " pod="openstack-operators/openstack-operator-index-vnslc" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.310717 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6grgs\" (UniqueName: \"kubernetes.io/projected/26967c39-74e8-4b89-89f8-e200c602de61-kube-api-access-6grgs\") pod \"openstack-operator-index-vnslc\" (UID: \"26967c39-74e8-4b89-89f8-e200c602de61\") " pod="openstack-operators/openstack-operator-index-vnslc" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.441762 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vnslc" Feb 19 08:57:12 crc kubenswrapper[4675]: I0219 08:57:12.848468 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vnslc"] Feb 19 08:57:13 crc kubenswrapper[4675]: I0219 08:57:13.212431 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:57:13 crc kubenswrapper[4675]: I0219 08:57:13.253038 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:57:13 crc kubenswrapper[4675]: I0219 08:57:13.343843 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vnslc" event={"ID":"26967c39-74e8-4b89-89f8-e200c602de61","Type":"ContainerStarted","Data":"e5fd4c0efe58d54842e55b08cf9f391588c4b396880e9fca3f028f8b6f2c9e84"} Feb 19 08:57:15 crc kubenswrapper[4675]: I0219 08:57:15.314037 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vnslc"] Feb 19 08:57:15 crc kubenswrapper[4675]: I0219 08:57:15.940151 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-h9qft"] Feb 19 08:57:15 crc kubenswrapper[4675]: I0219 08:57:15.941323 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:15 crc kubenswrapper[4675]: I0219 08:57:15.944030 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h9qft"] Feb 19 08:57:15 crc kubenswrapper[4675]: I0219 08:57:15.944752 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcdwg\" (UniqueName: \"kubernetes.io/projected/8aa0b598-0a09-4439-9e9d-c125fb5542db-kube-api-access-zcdwg\") pod \"openstack-operator-index-h9qft\" (UID: \"8aa0b598-0a09-4439-9e9d-c125fb5542db\") " pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.046339 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcdwg\" (UniqueName: \"kubernetes.io/projected/8aa0b598-0a09-4439-9e9d-c125fb5542db-kube-api-access-zcdwg\") pod \"openstack-operator-index-h9qft\" (UID: \"8aa0b598-0a09-4439-9e9d-c125fb5542db\") " pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.064705 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcdwg\" (UniqueName: \"kubernetes.io/projected/8aa0b598-0a09-4439-9e9d-c125fb5542db-kube-api-access-zcdwg\") pod \"openstack-operator-index-h9qft\" (UID: \"8aa0b598-0a09-4439-9e9d-c125fb5542db\") " pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.256617 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.370576 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vnslc" event={"ID":"26967c39-74e8-4b89-89f8-e200c602de61","Type":"ContainerStarted","Data":"ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e"} Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.370782 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-vnslc" podUID="26967c39-74e8-4b89-89f8-e200c602de61" containerName="registry-server" containerID="cri-o://ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e" gracePeriod=2 Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.392177 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vnslc" podStartSLOduration=1.488452428 podStartE2EDuration="4.392157904s" podCreationTimestamp="2026-02-19 08:57:12 +0000 UTC" firstStartedPulling="2026-02-19 08:57:12.856610941 +0000 UTC m=+814.483701219" lastFinishedPulling="2026-02-19 08:57:15.760316417 +0000 UTC m=+817.387406695" observedRunningTime="2026-02-19 08:57:16.389752318 +0000 UTC m=+818.016842586" watchObservedRunningTime="2026-02-19 08:57:16.392157904 +0000 UTC m=+818.019248172" Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.693301 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h9qft"] Feb 19 08:57:16 crc kubenswrapper[4675]: W0219 08:57:16.704204 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8aa0b598_0a09_4439_9e9d_c125fb5542db.slice/crio-1e156a042885eaca9ec6587d104e19bf4f5c6707473b4a1ea07fd7bb7be982e1 WatchSource:0}: Error finding container 1e156a042885eaca9ec6587d104e19bf4f5c6707473b4a1ea07fd7bb7be982e1: Status 404 returned error can't find the container with id 1e156a042885eaca9ec6587d104e19bf4f5c6707473b4a1ea07fd7bb7be982e1 Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.720845 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vnslc" Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.857605 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6grgs\" (UniqueName: \"kubernetes.io/projected/26967c39-74e8-4b89-89f8-e200c602de61-kube-api-access-6grgs\") pod \"26967c39-74e8-4b89-89f8-e200c602de61\" (UID: \"26967c39-74e8-4b89-89f8-e200c602de61\") " Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.863951 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26967c39-74e8-4b89-89f8-e200c602de61-kube-api-access-6grgs" (OuterVolumeSpecName: "kube-api-access-6grgs") pod "26967c39-74e8-4b89-89f8-e200c602de61" (UID: "26967c39-74e8-4b89-89f8-e200c602de61"). InnerVolumeSpecName "kube-api-access-6grgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:57:16 crc kubenswrapper[4675]: I0219 08:57:16.959299 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6grgs\" (UniqueName: \"kubernetes.io/projected/26967c39-74e8-4b89-89f8-e200c602de61-kube-api-access-6grgs\") on node \"crc\" DevicePath \"\"" Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.378068 4675 generic.go:334] "Generic (PLEG): container finished" podID="26967c39-74e8-4b89-89f8-e200c602de61" containerID="ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e" exitCode=0 Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.378127 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vnslc" Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.378130 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vnslc" event={"ID":"26967c39-74e8-4b89-89f8-e200c602de61","Type":"ContainerDied","Data":"ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e"} Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.378232 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vnslc" event={"ID":"26967c39-74e8-4b89-89f8-e200c602de61","Type":"ContainerDied","Data":"e5fd4c0efe58d54842e55b08cf9f391588c4b396880e9fca3f028f8b6f2c9e84"} Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.378248 4675 scope.go:117] "RemoveContainer" containerID="ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e" Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.379220 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h9qft" event={"ID":"8aa0b598-0a09-4439-9e9d-c125fb5542db","Type":"ContainerStarted","Data":"b1749a166b0d0aba0a098ac1c30617902a32454617f05923fd51ab80e93b4cc0"} Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.379238 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h9qft" event={"ID":"8aa0b598-0a09-4439-9e9d-c125fb5542db","Type":"ContainerStarted","Data":"1e156a042885eaca9ec6587d104e19bf4f5c6707473b4a1ea07fd7bb7be982e1"} Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.394796 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vnslc"] Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.398875 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-vnslc"] Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.400432 4675 scope.go:117] "RemoveContainer" containerID="ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e" Feb 19 08:57:17 crc kubenswrapper[4675]: E0219 08:57:17.400912 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e\": container with ID starting with ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e not found: ID does not exist" containerID="ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e" Feb 19 08:57:17 crc kubenswrapper[4675]: I0219 08:57:17.400950 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e"} err="failed to get container status \"ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e\": rpc error: code = NotFound desc = could not find container \"ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e\": container with ID starting with ae3d0c2833584f9044c1c282ed9c4e3205aa0c9dd680803355e192478c626b1e not found: ID does not exist" Feb 19 08:57:18 crc kubenswrapper[4675]: I0219 08:57:18.193850 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-fjr4t" Feb 19 08:57:18 crc kubenswrapper[4675]: I0219 08:57:18.211670 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-h9qft" podStartSLOduration=3.158715197 podStartE2EDuration="3.211627401s" podCreationTimestamp="2026-02-19 08:57:15 +0000 UTC" firstStartedPulling="2026-02-19 08:57:16.710074991 +0000 UTC m=+818.337165259" lastFinishedPulling="2026-02-19 08:57:16.762987185 +0000 UTC m=+818.390077463" observedRunningTime="2026-02-19 08:57:17.411919215 +0000 UTC m=+819.039009473" watchObservedRunningTime="2026-02-19 08:57:18.211627401 +0000 UTC m=+819.838717669" Feb 19 08:57:19 crc kubenswrapper[4675]: I0219 08:57:19.116195 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26967c39-74e8-4b89-89f8-e200c602de61" path="/var/lib/kubelet/pods/26967c39-74e8-4b89-89f8-e200c602de61/volumes" Feb 19 08:57:26 crc kubenswrapper[4675]: I0219 08:57:26.257031 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:26 crc kubenswrapper[4675]: I0219 08:57:26.257767 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:26 crc kubenswrapper[4675]: I0219 08:57:26.286901 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:26 crc kubenswrapper[4675]: I0219 08:57:26.473672 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-h9qft" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.351905 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl"] Feb 19 08:57:27 crc kubenswrapper[4675]: E0219 08:57:27.352379 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26967c39-74e8-4b89-89f8-e200c602de61" containerName="registry-server" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.352397 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="26967c39-74e8-4b89-89f8-e200c602de61" containerName="registry-server" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.352516 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="26967c39-74e8-4b89-89f8-e200c602de61" containerName="registry-server" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.353528 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.356507 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-ptq5h" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.361568 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl"] Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.393009 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-bundle\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.393094 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-util\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.393116 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcrrz\" (UniqueName: \"kubernetes.io/projected/f2fb95eb-0086-4762-96c1-fe6956a852e2-kube-api-access-qcrrz\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.494580 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-util\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.494627 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcrrz\" (UniqueName: \"kubernetes.io/projected/f2fb95eb-0086-4762-96c1-fe6956a852e2-kube-api-access-qcrrz\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.494726 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-bundle\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.495111 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-util\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.495122 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-bundle\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.514284 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcrrz\" (UniqueName: \"kubernetes.io/projected/f2fb95eb-0086-4762-96c1-fe6956a852e2-kube-api-access-qcrrz\") pod \"40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.672721 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:27 crc kubenswrapper[4675]: I0219 08:57:27.863335 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl"] Feb 19 08:57:27 crc kubenswrapper[4675]: W0219 08:57:27.869018 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2fb95eb_0086_4762_96c1_fe6956a852e2.slice/crio-b7c39090523b7a46ba3d1f68ab25ac9fc77bbe07d53ac7b37bf0de9297afa302 WatchSource:0}: Error finding container b7c39090523b7a46ba3d1f68ab25ac9fc77bbe07d53ac7b37bf0de9297afa302: Status 404 returned error can't find the container with id b7c39090523b7a46ba3d1f68ab25ac9fc77bbe07d53ac7b37bf0de9297afa302 Feb 19 08:57:28 crc kubenswrapper[4675]: I0219 08:57:28.213436 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-fj4cv" Feb 19 08:57:28 crc kubenswrapper[4675]: I0219 08:57:28.458345 4675 generic.go:334] "Generic (PLEG): container finished" podID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerID="ba6a67a064f3169cc1b57715c94a708850e31d64c6b73939ee920c80f56fa705" exitCode=0 Feb 19 08:57:28 crc kubenswrapper[4675]: I0219 08:57:28.458453 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" event={"ID":"f2fb95eb-0086-4762-96c1-fe6956a852e2","Type":"ContainerDied","Data":"ba6a67a064f3169cc1b57715c94a708850e31d64c6b73939ee920c80f56fa705"} Feb 19 08:57:28 crc kubenswrapper[4675]: I0219 08:57:28.458939 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" event={"ID":"f2fb95eb-0086-4762-96c1-fe6956a852e2","Type":"ContainerStarted","Data":"b7c39090523b7a46ba3d1f68ab25ac9fc77bbe07d53ac7b37bf0de9297afa302"} Feb 19 08:57:29 crc kubenswrapper[4675]: I0219 08:57:29.467276 4675 generic.go:334] "Generic (PLEG): container finished" podID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerID="4cf97ec8f0fe6607e61a62620fea7b7ac74760e30d7b8103a6498531cc79466c" exitCode=0 Feb 19 08:57:29 crc kubenswrapper[4675]: I0219 08:57:29.467337 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" event={"ID":"f2fb95eb-0086-4762-96c1-fe6956a852e2","Type":"ContainerDied","Data":"4cf97ec8f0fe6607e61a62620fea7b7ac74760e30d7b8103a6498531cc79466c"} Feb 19 08:57:30 crc kubenswrapper[4675]: I0219 08:57:30.480315 4675 generic.go:334] "Generic (PLEG): container finished" podID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerID="91599da24d23ec49531d21e513c95943caaa0c4ddb0912a70a3672e11db77a78" exitCode=0 Feb 19 08:57:30 crc kubenswrapper[4675]: I0219 08:57:30.480366 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" event={"ID":"f2fb95eb-0086-4762-96c1-fe6956a852e2","Type":"ContainerDied","Data":"91599da24d23ec49531d21e513c95943caaa0c4ddb0912a70a3672e11db77a78"} Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.735286 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.884476 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-bundle\") pod \"f2fb95eb-0086-4762-96c1-fe6956a852e2\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.884566 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcrrz\" (UniqueName: \"kubernetes.io/projected/f2fb95eb-0086-4762-96c1-fe6956a852e2-kube-api-access-qcrrz\") pod \"f2fb95eb-0086-4762-96c1-fe6956a852e2\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.884663 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-util\") pod \"f2fb95eb-0086-4762-96c1-fe6956a852e2\" (UID: \"f2fb95eb-0086-4762-96c1-fe6956a852e2\") " Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.885498 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-bundle" (OuterVolumeSpecName: "bundle") pod "f2fb95eb-0086-4762-96c1-fe6956a852e2" (UID: "f2fb95eb-0086-4762-96c1-fe6956a852e2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.890474 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2fb95eb-0086-4762-96c1-fe6956a852e2-kube-api-access-qcrrz" (OuterVolumeSpecName: "kube-api-access-qcrrz") pod "f2fb95eb-0086-4762-96c1-fe6956a852e2" (UID: "f2fb95eb-0086-4762-96c1-fe6956a852e2"). InnerVolumeSpecName "kube-api-access-qcrrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.899448 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-util" (OuterVolumeSpecName: "util") pod "f2fb95eb-0086-4762-96c1-fe6956a852e2" (UID: "f2fb95eb-0086-4762-96c1-fe6956a852e2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.986131 4675 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.986168 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcrrz\" (UniqueName: \"kubernetes.io/projected/f2fb95eb-0086-4762-96c1-fe6956a852e2-kube-api-access-qcrrz\") on node \"crc\" DevicePath \"\"" Feb 19 08:57:31 crc kubenswrapper[4675]: I0219 08:57:31.986179 4675 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f2fb95eb-0086-4762-96c1-fe6956a852e2-util\") on node \"crc\" DevicePath \"\"" Feb 19 08:57:32 crc kubenswrapper[4675]: I0219 08:57:32.492127 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" event={"ID":"f2fb95eb-0086-4762-96c1-fe6956a852e2","Type":"ContainerDied","Data":"b7c39090523b7a46ba3d1f68ab25ac9fc77bbe07d53ac7b37bf0de9297afa302"} Feb 19 08:57:32 crc kubenswrapper[4675]: I0219 08:57:32.492166 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7c39090523b7a46ba3d1f68ab25ac9fc77bbe07d53ac7b37bf0de9297afa302" Feb 19 08:57:32 crc kubenswrapper[4675]: I0219 08:57:32.492180 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.700274 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82"] Feb 19 08:57:39 crc kubenswrapper[4675]: E0219 08:57:39.701262 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerName="util" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.701277 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerName="util" Feb 19 08:57:39 crc kubenswrapper[4675]: E0219 08:57:39.701288 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerName="pull" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.701296 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerName="pull" Feb 19 08:57:39 crc kubenswrapper[4675]: E0219 08:57:39.701315 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerName="extract" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.701324 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerName="extract" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.701457 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2fb95eb-0086-4762-96c1-fe6956a852e2" containerName="extract" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.702027 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.704178 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-zrmwm" Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.721706 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82"] Feb 19 08:57:39 crc kubenswrapper[4675]: I0219 08:57:39.901512 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpqtg\" (UniqueName: \"kubernetes.io/projected/f7d62133-059d-4695-829f-ed7b9888e323-kube-api-access-fpqtg\") pod \"openstack-operator-controller-init-767d85d5f6-7tx82\" (UID: \"f7d62133-059d-4695-829f-ed7b9888e323\") " pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" Feb 19 08:57:40 crc kubenswrapper[4675]: I0219 08:57:40.002716 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpqtg\" (UniqueName: \"kubernetes.io/projected/f7d62133-059d-4695-829f-ed7b9888e323-kube-api-access-fpqtg\") pod \"openstack-operator-controller-init-767d85d5f6-7tx82\" (UID: \"f7d62133-059d-4695-829f-ed7b9888e323\") " pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" Feb 19 08:57:40 crc kubenswrapper[4675]: I0219 08:57:40.022261 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpqtg\" (UniqueName: \"kubernetes.io/projected/f7d62133-059d-4695-829f-ed7b9888e323-kube-api-access-fpqtg\") pod \"openstack-operator-controller-init-767d85d5f6-7tx82\" (UID: \"f7d62133-059d-4695-829f-ed7b9888e323\") " pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" Feb 19 08:57:40 crc kubenswrapper[4675]: I0219 08:57:40.318930 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" Feb 19 08:57:40 crc kubenswrapper[4675]: I0219 08:57:40.742148 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82"] Feb 19 08:57:40 crc kubenswrapper[4675]: W0219 08:57:40.747731 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7d62133_059d_4695_829f_ed7b9888e323.slice/crio-2c791698767ade7477e1f2dc593413c256dbde3ca50a1097da502c41e24f86a5 WatchSource:0}: Error finding container 2c791698767ade7477e1f2dc593413c256dbde3ca50a1097da502c41e24f86a5: Status 404 returned error can't find the container with id 2c791698767ade7477e1f2dc593413c256dbde3ca50a1097da502c41e24f86a5 Feb 19 08:57:41 crc kubenswrapper[4675]: I0219 08:57:41.549847 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" event={"ID":"f7d62133-059d-4695-829f-ed7b9888e323","Type":"ContainerStarted","Data":"2c791698767ade7477e1f2dc593413c256dbde3ca50a1097da502c41e24f86a5"} Feb 19 08:57:41 crc kubenswrapper[4675]: I0219 08:57:41.743625 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 08:57:41 crc kubenswrapper[4675]: I0219 08:57:41.743715 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 08:57:41 crc kubenswrapper[4675]: I0219 08:57:41.743773 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 08:57:41 crc kubenswrapper[4675]: I0219 08:57:41.744419 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ee4c969741bfa1ae0382fcd0dac3640f8d7616539fc52b4655b88e929fd6acf"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 08:57:41 crc kubenswrapper[4675]: I0219 08:57:41.744520 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://8ee4c969741bfa1ae0382fcd0dac3640f8d7616539fc52b4655b88e929fd6acf" gracePeriod=600 Feb 19 08:57:42 crc kubenswrapper[4675]: I0219 08:57:42.558742 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="8ee4c969741bfa1ae0382fcd0dac3640f8d7616539fc52b4655b88e929fd6acf" exitCode=0 Feb 19 08:57:42 crc kubenswrapper[4675]: I0219 08:57:42.558786 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"8ee4c969741bfa1ae0382fcd0dac3640f8d7616539fc52b4655b88e929fd6acf"} Feb 19 08:57:42 crc kubenswrapper[4675]: I0219 08:57:42.558819 4675 scope.go:117] "RemoveContainer" containerID="011bd75815a488feaac53d8c8b5eb25ad12052f61153f59a628707c37551dfbb" Feb 19 08:57:45 crc kubenswrapper[4675]: I0219 08:57:45.583692 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" event={"ID":"f7d62133-059d-4695-829f-ed7b9888e323","Type":"ContainerStarted","Data":"1b300c4d403e79fdae17ae898e30085454e6d52c2db677d5f9662adf776a136a"} Feb 19 08:57:45 crc kubenswrapper[4675]: I0219 08:57:45.584395 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" Feb 19 08:57:45 crc kubenswrapper[4675]: I0219 08:57:45.586130 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"ca5833a715b96b50d8038acbecde68c582755da122b5dea38657e570ff140d31"} Feb 19 08:57:45 crc kubenswrapper[4675]: I0219 08:57:45.656272 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" podStartSLOduration=2.347872745 podStartE2EDuration="6.65625314s" podCreationTimestamp="2026-02-19 08:57:39 +0000 UTC" firstStartedPulling="2026-02-19 08:57:40.750664022 +0000 UTC m=+842.377754290" lastFinishedPulling="2026-02-19 08:57:45.059044417 +0000 UTC m=+846.686134685" observedRunningTime="2026-02-19 08:57:45.654599005 +0000 UTC m=+847.281689283" watchObservedRunningTime="2026-02-19 08:57:45.65625314 +0000 UTC m=+847.283343408" Feb 19 08:57:50 crc kubenswrapper[4675]: I0219 08:57:50.322921 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-767d85d5f6-7tx82" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.580376 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.587542 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.587542 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.588203 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.589948 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-9529c" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.590209 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cms2p" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.602746 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.607453 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.627538 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.628433 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.633122 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.634398 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.638855 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2l6d9" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.639584 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-g52s4" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.641460 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.657309 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.663100 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.663893 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.670043 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-w8rc6" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.685633 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.699772 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.700755 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.705180 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-cpmdc" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.719337 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9j7w\" (UniqueName: \"kubernetes.io/projected/2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282-kube-api-access-b9j7w\") pod \"barbican-operator-controller-manager-868647ff47-bj7kj\" (UID: \"2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.719371 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krnfg\" (UniqueName: \"kubernetes.io/projected/f79e0726-e5c1-4536-b06b-40bab3849bb0-kube-api-access-krnfg\") pod \"cinder-operator-controller-manager-5d946d989d-c775s\" (UID: \"f79e0726-e5c1-4536-b06b-40bab3849bb0\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.730344 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.731322 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.735872 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.736049 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-twvvd" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.738553 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.750557 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.755602 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.759064 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.762077 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-dlh4q" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.764094 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.765050 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.768379 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vvm8s" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.772257 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.807756 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.827228 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67hrq\" (UniqueName: \"kubernetes.io/projected/53318e37-b20a-46eb-8f12-decb087ade91-kube-api-access-67hrq\") pod \"heat-operator-controller-manager-69f49c598c-bwwwp\" (UID: \"53318e37-b20a-46eb-8f12-decb087ade91\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.827284 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncmct\" (UniqueName: \"kubernetes.io/projected/b0de8493-3f97-4e5e-a28d-5d45f545d645-kube-api-access-ncmct\") pod \"glance-operator-controller-manager-77987464f4-v7lcl\" (UID: \"b0de8493-3f97-4e5e-a28d-5d45f545d645\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.827329 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9j7w\" (UniqueName: \"kubernetes.io/projected/2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282-kube-api-access-b9j7w\") pod \"barbican-operator-controller-manager-868647ff47-bj7kj\" (UID: \"2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.827956 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krnfg\" (UniqueName: \"kubernetes.io/projected/f79e0726-e5c1-4536-b06b-40bab3849bb0-kube-api-access-krnfg\") pod \"cinder-operator-controller-manager-5d946d989d-c775s\" (UID: \"f79e0726-e5c1-4536-b06b-40bab3849bb0\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.828029 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fr28\" (UniqueName: \"kubernetes.io/projected/7f7fd1c2-d277-402d-a762-0ebfddc48226-kube-api-access-4fr28\") pod \"designate-operator-controller-manager-6d8bf5c495-glh49\" (UID: \"7f7fd1c2-d277-402d-a762-0ebfddc48226\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.828056 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knvkp\" (UniqueName: \"kubernetes.io/projected/2473d2e8-7c4d-41f5-bd2a-95823fe059f1-kube-api-access-knvkp\") pod \"horizon-operator-controller-manager-5b9b8895d5-9zn82\" (UID: \"2473d2e8-7c4d-41f5-bd2a-95823fe059f1\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.841081 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.848033 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.857133 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-4mvf6" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.875377 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.903341 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.903884 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9j7w\" (UniqueName: \"kubernetes.io/projected/2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282-kube-api-access-b9j7w\") pod \"barbican-operator-controller-manager-868647ff47-bj7kj\" (UID: \"2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.908976 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krnfg\" (UniqueName: \"kubernetes.io/projected/f79e0726-e5c1-4536-b06b-40bab3849bb0-kube-api-access-krnfg\") pod \"cinder-operator-controller-manager-5d946d989d-c775s\" (UID: \"f79e0726-e5c1-4536-b06b-40bab3849bb0\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.911092 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.916514 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-c94lz" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.921359 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.921825 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929319 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67hrq\" (UniqueName: \"kubernetes.io/projected/53318e37-b20a-46eb-8f12-decb087ade91-kube-api-access-67hrq\") pod \"heat-operator-controller-manager-69f49c598c-bwwwp\" (UID: \"53318e37-b20a-46eb-8f12-decb087ade91\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929384 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncmct\" (UniqueName: \"kubernetes.io/projected/b0de8493-3f97-4e5e-a28d-5d45f545d645-kube-api-access-ncmct\") pod \"glance-operator-controller-manager-77987464f4-v7lcl\" (UID: \"b0de8493-3f97-4e5e-a28d-5d45f545d645\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929414 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929439 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c52c8\" (UniqueName: \"kubernetes.io/projected/78209850-b313-4190-a3a4-674da1afaaba-kube-api-access-c52c8\") pod \"ironic-operator-controller-manager-554564d7fc-z2wq2\" (UID: \"78209850-b313-4190-a3a4-674da1afaaba\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929502 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trbn9\" (UniqueName: \"kubernetes.io/projected/4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6-kube-api-access-trbn9\") pod \"keystone-operator-controller-manager-b4d948c87-jhgw4\" (UID: \"4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929532 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fr28\" (UniqueName: \"kubernetes.io/projected/7f7fd1c2-d277-402d-a762-0ebfddc48226-kube-api-access-4fr28\") pod \"designate-operator-controller-manager-6d8bf5c495-glh49\" (UID: \"7f7fd1c2-d277-402d-a762-0ebfddc48226\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929553 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knvkp\" (UniqueName: \"kubernetes.io/projected/2473d2e8-7c4d-41f5-bd2a-95823fe059f1-kube-api-access-knvkp\") pod \"horizon-operator-controller-manager-5b9b8895d5-9zn82\" (UID: \"2473d2e8-7c4d-41f5-bd2a-95823fe059f1\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.929595 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l5r6\" (UniqueName: \"kubernetes.io/projected/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-kube-api-access-8l5r6\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.930782 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.931546 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.934581 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-c2dg5" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.947041 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.953425 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fr28\" (UniqueName: \"kubernetes.io/projected/7f7fd1c2-d277-402d-a762-0ebfddc48226-kube-api-access-4fr28\") pod \"designate-operator-controller-manager-6d8bf5c495-glh49\" (UID: \"7f7fd1c2-d277-402d-a762-0ebfddc48226\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.955370 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncmct\" (UniqueName: \"kubernetes.io/projected/b0de8493-3f97-4e5e-a28d-5d45f545d645-kube-api-access-ncmct\") pod \"glance-operator-controller-manager-77987464f4-v7lcl\" (UID: \"b0de8493-3f97-4e5e-a28d-5d45f545d645\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.955537 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.956390 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67hrq\" (UniqueName: \"kubernetes.io/projected/53318e37-b20a-46eb-8f12-decb087ade91-kube-api-access-67hrq\") pod \"heat-operator-controller-manager-69f49c598c-bwwwp\" (UID: \"53318e37-b20a-46eb-8f12-decb087ade91\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.958040 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knvkp\" (UniqueName: \"kubernetes.io/projected/2473d2e8-7c4d-41f5-bd2a-95823fe059f1-kube-api-access-knvkp\") pod \"horizon-operator-controller-manager-5b9b8895d5-9zn82\" (UID: \"2473d2e8-7c4d-41f5-bd2a-95823fe059f1\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.958289 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.963717 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.965703 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.968773 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-9c6fk" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.969545 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.970490 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.971040 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.972002 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-lwxtq" Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.978304 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.986038 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.994221 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8"] Feb 19 08:58:09 crc kubenswrapper[4675]: I0219 08:58:09.998091 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.000955 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.001297 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-h9pmt" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.001428 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.002375 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.003229 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.008412 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.009330 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.014209 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-tlv95" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.014406 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-bps8h" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.019520 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.024011 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.030947 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7kfl\" (UniqueName: \"kubernetes.io/projected/63fcb9d8-c313-4d0e-9f53-44c3ad272d1d-kube-api-access-n7kfl\") pod \"manila-operator-controller-manager-54f6768c69-kqz2s\" (UID: \"63fcb9d8-c313-4d0e-9f53-44c3ad272d1d\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.030985 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sd8s\" (UniqueName: \"kubernetes.io/projected/61877abc-2f6b-40a5-aa19-b8adf5402cc6-kube-api-access-4sd8s\") pod \"mariadb-operator-controller-manager-6994f66f48-c25kr\" (UID: \"61877abc-2f6b-40a5-aa19-b8adf5402cc6\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.031018 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.031041 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c52c8\" (UniqueName: \"kubernetes.io/projected/78209850-b313-4190-a3a4-674da1afaaba-kube-api-access-c52c8\") pod \"ironic-operator-controller-manager-554564d7fc-z2wq2\" (UID: \"78209850-b313-4190-a3a4-674da1afaaba\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.031094 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trbn9\" (UniqueName: \"kubernetes.io/projected/4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6-kube-api-access-trbn9\") pod \"keystone-operator-controller-manager-b4d948c87-jhgw4\" (UID: \"4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.031131 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l5r6\" (UniqueName: \"kubernetes.io/projected/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-kube-api-access-8l5r6\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.031465 4675 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.031512 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert podName:a5bfbb7f-0d4a-49d3-bf77-b168ccf33061 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:10.531496937 +0000 UTC m=+872.158587205 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert") pod "infra-operator-controller-manager-79d975b745-cbbkp" (UID: "a5bfbb7f-0d4a-49d3-bf77-b168ccf33061") : secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.039743 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.051501 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trbn9\" (UniqueName: \"kubernetes.io/projected/4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6-kube-api-access-trbn9\") pod \"keystone-operator-controller-manager-b4d948c87-jhgw4\" (UID: \"4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.051541 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l5r6\" (UniqueName: \"kubernetes.io/projected/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-kube-api-access-8l5r6\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.051790 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.058667 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c52c8\" (UniqueName: \"kubernetes.io/projected/78209850-b313-4190-a3a4-674da1afaaba-kube-api-access-c52c8\") pod \"ironic-operator-controller-manager-554564d7fc-z2wq2\" (UID: \"78209850-b313-4190-a3a4-674da1afaaba\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.061255 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.062111 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.069835 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.071328 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-vx5j8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.098691 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.099407 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.100302 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.103065 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.106078 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-2mgf8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.107537 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.134927 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-w8w5t"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.135809 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137248 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc65s\" (UniqueName: \"kubernetes.io/projected/70a13bde-19ab-4c7c-af9a-15f1fe96d4e1-kube-api-access-bc65s\") pod \"test-operator-controller-manager-7866795846-w8w5t\" (UID: \"70a13bde-19ab-4c7c-af9a-15f1fe96d4e1\") " pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137270 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q4k2\" (UniqueName: \"kubernetes.io/projected/53989876-5869-4e41-a9c3-c99cea8347bd-kube-api-access-7q4k2\") pod \"swift-operator-controller-manager-68f46476f-jz6fw\" (UID: \"53989876-5869-4e41-a9c3-c99cea8347bd\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137311 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7kfl\" (UniqueName: \"kubernetes.io/projected/63fcb9d8-c313-4d0e-9f53-44c3ad272d1d-kube-api-access-n7kfl\") pod \"manila-operator-controller-manager-54f6768c69-kqz2s\" (UID: \"63fcb9d8-c313-4d0e-9f53-44c3ad272d1d\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137332 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sd8s\" (UniqueName: \"kubernetes.io/projected/61877abc-2f6b-40a5-aa19-b8adf5402cc6-kube-api-access-4sd8s\") pod \"mariadb-operator-controller-manager-6994f66f48-c25kr\" (UID: \"61877abc-2f6b-40a5-aa19-b8adf5402cc6\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137372 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwk48\" (UniqueName: \"kubernetes.io/projected/2f4e8661-38a2-44bc-90ae-266f058758a9-kube-api-access-jwk48\") pod \"placement-operator-controller-manager-8497b45c89-dr2nj\" (UID: \"2f4e8661-38a2-44bc-90ae-266f058758a9\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137394 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jghnv\" (UniqueName: \"kubernetes.io/projected/4fad9fc9-cb88-4aff-89bb-0a2fd609f48a-kube-api-access-jghnv\") pod \"octavia-operator-controller-manager-69f8888797-ff4m9\" (UID: \"4fad9fc9-cb88-4aff-89bb-0a2fd609f48a\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137412 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpqg6\" (UniqueName: \"kubernetes.io/projected/4e163d5d-3666-4899-acd7-21040c1cc573-kube-api-access-bpqg6\") pod \"nova-operator-controller-manager-567668f5cf-ddhjt\" (UID: \"4e163d5d-3666-4899-acd7-21040c1cc573\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137439 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137456 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwxn\" (UniqueName: \"kubernetes.io/projected/5c9124b1-4ca0-481f-8474-ea45252bda76-kube-api-access-mgwxn\") pod \"neutron-operator-controller-manager-64ddbf8bb-8k6xs\" (UID: \"5c9124b1-4ca0-481f-8474-ea45252bda76\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137480 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpfk2\" (UniqueName: \"kubernetes.io/projected/b1bbb05f-c65b-4ea1-ab19-34b888f49369-kube-api-access-tpfk2\") pod \"telemetry-operator-controller-manager-7f45b4ff68-s25g8\" (UID: \"b1bbb05f-c65b-4ea1-ab19-34b888f49369\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137495 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kgrh\" (UniqueName: \"kubernetes.io/projected/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-kube-api-access-4kgrh\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.137517 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvwhl\" (UniqueName: \"kubernetes.io/projected/dcd57b86-afb5-4beb-ae73-830ce75a562f-kube-api-access-nvwhl\") pod \"ovn-operator-controller-manager-d44cf6b75-r8bkh\" (UID: \"dcd57b86-afb5-4beb-ae73-830ce75a562f\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.138108 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-zdq6l" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.148695 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-w8w5t"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.164701 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sd8s\" (UniqueName: \"kubernetes.io/projected/61877abc-2f6b-40a5-aa19-b8adf5402cc6-kube-api-access-4sd8s\") pod \"mariadb-operator-controller-manager-6994f66f48-c25kr\" (UID: \"61877abc-2f6b-40a5-aa19-b8adf5402cc6\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.179337 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7kfl\" (UniqueName: \"kubernetes.io/projected/63fcb9d8-c313-4d0e-9f53-44c3ad272d1d-kube-api-access-n7kfl\") pod \"manila-operator-controller-manager-54f6768c69-kqz2s\" (UID: \"63fcb9d8-c313-4d0e-9f53-44c3ad272d1d\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.211217 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238178 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwk48\" (UniqueName: \"kubernetes.io/projected/2f4e8661-38a2-44bc-90ae-266f058758a9-kube-api-access-jwk48\") pod \"placement-operator-controller-manager-8497b45c89-dr2nj\" (UID: \"2f4e8661-38a2-44bc-90ae-266f058758a9\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238226 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jghnv\" (UniqueName: \"kubernetes.io/projected/4fad9fc9-cb88-4aff-89bb-0a2fd609f48a-kube-api-access-jghnv\") pod \"octavia-operator-controller-manager-69f8888797-ff4m9\" (UID: \"4fad9fc9-cb88-4aff-89bb-0a2fd609f48a\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238249 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpqg6\" (UniqueName: \"kubernetes.io/projected/4e163d5d-3666-4899-acd7-21040c1cc573-kube-api-access-bpqg6\") pod \"nova-operator-controller-manager-567668f5cf-ddhjt\" (UID: \"4e163d5d-3666-4899-acd7-21040c1cc573\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238274 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238296 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwxn\" (UniqueName: \"kubernetes.io/projected/5c9124b1-4ca0-481f-8474-ea45252bda76-kube-api-access-mgwxn\") pod \"neutron-operator-controller-manager-64ddbf8bb-8k6xs\" (UID: \"5c9124b1-4ca0-481f-8474-ea45252bda76\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238322 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpfk2\" (UniqueName: \"kubernetes.io/projected/b1bbb05f-c65b-4ea1-ab19-34b888f49369-kube-api-access-tpfk2\") pod \"telemetry-operator-controller-manager-7f45b4ff68-s25g8\" (UID: \"b1bbb05f-c65b-4ea1-ab19-34b888f49369\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238340 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kgrh\" (UniqueName: \"kubernetes.io/projected/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-kube-api-access-4kgrh\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238360 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvwhl\" (UniqueName: \"kubernetes.io/projected/dcd57b86-afb5-4beb-ae73-830ce75a562f-kube-api-access-nvwhl\") pod \"ovn-operator-controller-manager-d44cf6b75-r8bkh\" (UID: \"dcd57b86-afb5-4beb-ae73-830ce75a562f\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238379 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc65s\" (UniqueName: \"kubernetes.io/projected/70a13bde-19ab-4c7c-af9a-15f1fe96d4e1-kube-api-access-bc65s\") pod \"test-operator-controller-manager-7866795846-w8w5t\" (UID: \"70a13bde-19ab-4c7c-af9a-15f1fe96d4e1\") " pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.238397 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q4k2\" (UniqueName: \"kubernetes.io/projected/53989876-5869-4e41-a9c3-c99cea8347bd-kube-api-access-7q4k2\") pod \"swift-operator-controller-manager-68f46476f-jz6fw\" (UID: \"53989876-5869-4e41-a9c3-c99cea8347bd\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.239590 4675 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.239667 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert podName:769e3e7a-a9ea-4d9c-b60b-8643e0c275e1 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:10.739651301 +0000 UTC m=+872.366741569 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" (UID: "769e3e7a-a9ea-4d9c-b60b-8643e0c275e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.296962 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpfk2\" (UniqueName: \"kubernetes.io/projected/b1bbb05f-c65b-4ea1-ab19-34b888f49369-kube-api-access-tpfk2\") pod \"telemetry-operator-controller-manager-7f45b4ff68-s25g8\" (UID: \"b1bbb05f-c65b-4ea1-ab19-34b888f49369\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.297906 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpqg6\" (UniqueName: \"kubernetes.io/projected/4e163d5d-3666-4899-acd7-21040c1cc573-kube-api-access-bpqg6\") pod \"nova-operator-controller-manager-567668f5cf-ddhjt\" (UID: \"4e163d5d-3666-4899-acd7-21040c1cc573\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.301750 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc65s\" (UniqueName: \"kubernetes.io/projected/70a13bde-19ab-4c7c-af9a-15f1fe96d4e1-kube-api-access-bc65s\") pod \"test-operator-controller-manager-7866795846-w8w5t\" (UID: \"70a13bde-19ab-4c7c-af9a-15f1fe96d4e1\") " pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.306492 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.310245 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvwhl\" (UniqueName: \"kubernetes.io/projected/dcd57b86-afb5-4beb-ae73-830ce75a562f-kube-api-access-nvwhl\") pod \"ovn-operator-controller-manager-d44cf6b75-r8bkh\" (UID: \"dcd57b86-afb5-4beb-ae73-830ce75a562f\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.310747 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kgrh\" (UniqueName: \"kubernetes.io/projected/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-kube-api-access-4kgrh\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.312900 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwk48\" (UniqueName: \"kubernetes.io/projected/2f4e8661-38a2-44bc-90ae-266f058758a9-kube-api-access-jwk48\") pod \"placement-operator-controller-manager-8497b45c89-dr2nj\" (UID: \"2f4e8661-38a2-44bc-90ae-266f058758a9\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.328849 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.339869 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwxn\" (UniqueName: \"kubernetes.io/projected/5c9124b1-4ca0-481f-8474-ea45252bda76-kube-api-access-mgwxn\") pod \"neutron-operator-controller-manager-64ddbf8bb-8k6xs\" (UID: \"5c9124b1-4ca0-481f-8474-ea45252bda76\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.340660 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.341159 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px8qf\" (UniqueName: \"kubernetes.io/projected/1a383e65-12c2-43bf-b6b0-f61f20706d7a-kube-api-access-px8qf\") pod \"watcher-operator-controller-manager-5db88f68c-x6glz\" (UID: \"1a383e65-12c2-43bf-b6b0-f61f20706d7a\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.342948 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jghnv\" (UniqueName: \"kubernetes.io/projected/4fad9fc9-cb88-4aff-89bb-0a2fd609f48a-kube-api-access-jghnv\") pod \"octavia-operator-controller-manager-69f8888797-ff4m9\" (UID: \"4fad9fc9-cb88-4aff-89bb-0a2fd609f48a\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.344654 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q4k2\" (UniqueName: \"kubernetes.io/projected/53989876-5869-4e41-a9c3-c99cea8347bd-kube-api-access-7q4k2\") pod \"swift-operator-controller-manager-68f46476f-jz6fw\" (UID: \"53989876-5869-4e41-a9c3-c99cea8347bd\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.344795 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.346700 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2tbkq" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.355710 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.363325 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.384973 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.392348 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.407167 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.411742 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.412000 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.412017 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.412118 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-jwl72" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.427536 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.439110 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.443512 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.443549 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.443611 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px8qf\" (UniqueName: \"kubernetes.io/projected/1a383e65-12c2-43bf-b6b0-f61f20706d7a-kube-api-access-px8qf\") pod \"watcher-operator-controller-manager-5db88f68c-x6glz\" (UID: \"1a383e65-12c2-43bf-b6b0-f61f20706d7a\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.443677 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/a4559dd6-5152-40d8-94af-f2fff6c90e84-kube-api-access-mpczd\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.481679 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.483034 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.487171 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn"] Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.489535 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px8qf\" (UniqueName: \"kubernetes.io/projected/1a383e65-12c2-43bf-b6b0-f61f20706d7a-kube-api-access-px8qf\") pod \"watcher-operator-controller-manager-5db88f68c-x6glz\" (UID: \"1a383e65-12c2-43bf-b6b0-f61f20706d7a\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.502427 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-xjc59" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.518119 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.528571 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.544305 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.544340 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.544383 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.544431 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvndm\" (UniqueName: \"kubernetes.io/projected/b9dc6b16-af6b-470e-a9a7-d59245e8a22e-kube-api-access-mvndm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-sm2fn\" (UID: \"b9dc6b16-af6b-470e-a9a7-d59245e8a22e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.544464 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/a4559dd6-5152-40d8-94af-f2fff6c90e84-kube-api-access-mpczd\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.544839 4675 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.544882 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:11.044866992 +0000 UTC m=+872.671957260 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "metrics-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.545051 4675 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.545077 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert podName:a5bfbb7f-0d4a-49d3-bf77-b168ccf33061 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:11.545069888 +0000 UTC m=+873.172160156 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert") pod "infra-operator-controller-manager-79d975b745-cbbkp" (UID: "a5bfbb7f-0d4a-49d3-bf77-b168ccf33061") : secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.545202 4675 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.545292 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:11.045271183 +0000 UTC m=+872.672361451 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.565373 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/a4559dd6-5152-40d8-94af-f2fff6c90e84-kube-api-access-mpczd\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.588505 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.645744 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvndm\" (UniqueName: \"kubernetes.io/projected/b9dc6b16-af6b-470e-a9a7-d59245e8a22e-kube-api-access-mvndm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-sm2fn\" (UID: \"b9dc6b16-af6b-470e-a9a7-d59245e8a22e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.677575 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvndm\" (UniqueName: \"kubernetes.io/projected/b9dc6b16-af6b-470e-a9a7-d59245e8a22e-kube-api-access-mvndm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-sm2fn\" (UID: \"b9dc6b16-af6b-470e-a9a7-d59245e8a22e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.744066 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.746867 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.747362 4675 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: E0219 08:58:10.747416 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert podName:769e3e7a-a9ea-4d9c-b60b-8643e0c275e1 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:11.747401444 +0000 UTC m=+873.374491712 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" (UID: "769e3e7a-a9ea-4d9c-b60b-8643e0c275e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:10 crc kubenswrapper[4675]: I0219 08:58:10.934908 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.034755 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.037867 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.050737 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.050766 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.050938 4675 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.050966 4675 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.050988 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:12.050974501 +0000 UTC m=+873.678064769 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "metrics-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.051059 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:12.051035322 +0000 UTC m=+873.678125590 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "webhook-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.068418 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf79e0726_e5c1_4536_b06b_40bab3849bb0.slice/crio-992708b4b39de5935cb45337253f158dde574e8ced61841c61e0c217ff5dd0bb WatchSource:0}: Error finding container 992708b4b39de5935cb45337253f158dde574e8ced61841c61e0c217ff5dd0bb: Status 404 returned error can't find the container with id 992708b4b39de5935cb45337253f158dde574e8ced61841c61e0c217ff5dd0bb Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.413451 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.426195 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.433609 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.448989 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.469125 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.475845 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49"] Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.481380 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0de8493_3f97_4e5e_a28d_5d45f545d645.slice/crio-9e64ec5a7215a44e96188e7ee3f7ba7a033223793f3f9cf9f607c0b88e060cf6 WatchSource:0}: Error finding container 9e64ec5a7215a44e96188e7ee3f7ba7a033223793f3f9cf9f607c0b88e060cf6: Status 404 returned error can't find the container with id 9e64ec5a7215a44e96188e7ee3f7ba7a033223793f3f9cf9f607c0b88e060cf6 Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.552423 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.562537 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs"] Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.565051 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61877abc_2f6b_40a5_aa19_b8adf5402cc6.slice/crio-3603e59506c5d7bdcc8ece03da6ef16acf9fa428469ff93c951e75b05b58d82c WatchSource:0}: Error finding container 3603e59506c5d7bdcc8ece03da6ef16acf9fa428469ff93c951e75b05b58d82c: Status 404 returned error can't find the container with id 3603e59506c5d7bdcc8ece03da6ef16acf9fa428469ff93c951e75b05b58d82c Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.565794 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.566387 4675 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.566453 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert podName:a5bfbb7f-0d4a-49d3-bf77-b168ccf33061 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:13.566434963 +0000 UTC m=+875.193525231 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert") pod "infra-operator-controller-manager-79d975b745-cbbkp" (UID: "a5bfbb7f-0d4a-49d3-bf77-b168ccf33061") : secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.572738 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.577311 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.589856 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.590496 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.598387 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8"] Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.647743 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e163d5d_3666_4899_acd7_21040c1cc573.slice/crio-34d6d333851b5795e184cd36a71794f202fe10c6d17cb2710cc4b433d0422483 WatchSource:0}: Error finding container 34d6d333851b5795e184cd36a71794f202fe10c6d17cb2710cc4b433d0422483: Status 404 returned error can't find the container with id 34d6d333851b5795e184cd36a71794f202fe10c6d17cb2710cc4b433d0422483 Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.649011 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nvwhl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-d44cf6b75-r8bkh_openstack-operators(dcd57b86-afb5-4beb-ae73-830ce75a562f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.650593 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" podUID="dcd57b86-afb5-4beb-ae73-830ce75a562f" Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.668108 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fcb9d8_c313_4d0e_9f53_44c3ad272d1d.slice/crio-eacee856281625e9e41b5e50ed08741032ab394bc2c38a13e22bed4ad329e65a WatchSource:0}: Error finding container eacee856281625e9e41b5e50ed08741032ab394bc2c38a13e22bed4ad329e65a: Status 404 returned error can't find the container with id eacee856281625e9e41b5e50ed08741032ab394bc2c38a13e22bed4ad329e65a Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.675628 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:8fb0a33b8d93cf9f84f079af5f2ceb680afada4e44542514959146779f57f64c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n7kfl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-54f6768c69-kqz2s_openstack-operators(63fcb9d8-c313-4d0e-9f53-44c3ad272d1d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.683756 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" podUID="63fcb9d8-c313-4d0e-9f53-44c3ad272d1d" Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.770551 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.770880 4675 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.771059 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert podName:769e3e7a-a9ea-4d9c-b60b-8643e0c275e1 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:13.771038461 +0000 UTC m=+875.398128729 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" (UID: "769e3e7a-a9ea-4d9c-b60b-8643e0c275e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.784299 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" event={"ID":"f79e0726-e5c1-4536-b06b-40bab3849bb0","Type":"ContainerStarted","Data":"992708b4b39de5935cb45337253f158dde574e8ced61841c61e0c217ff5dd0bb"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.785317 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" event={"ID":"4e163d5d-3666-4899-acd7-21040c1cc573","Type":"ContainerStarted","Data":"34d6d333851b5795e184cd36a71794f202fe10c6d17cb2710cc4b433d0422483"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.787439 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" event={"ID":"4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6","Type":"ContainerStarted","Data":"bac53feb3b8645999d19528ae674707ad88c6a6c975aceae45497a1668322476"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.789415 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" event={"ID":"78209850-b313-4190-a3a4-674da1afaaba","Type":"ContainerStarted","Data":"3b002f9325bebeadfd8cac00c11685c4b58caf2d0e24029e23eacbdf92e584a7"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.793172 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" event={"ID":"63fcb9d8-c313-4d0e-9f53-44c3ad272d1d","Type":"ContainerStarted","Data":"eacee856281625e9e41b5e50ed08741032ab394bc2c38a13e22bed4ad329e65a"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.797453 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" event={"ID":"7f7fd1c2-d277-402d-a762-0ebfddc48226","Type":"ContainerStarted","Data":"c460f0a16913778a0e11703a68b44b65112140d19db85c33daad8d1fd134c13f"} Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.797486 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8fb0a33b8d93cf9f84f079af5f2ceb680afada4e44542514959146779f57f64c\\\"\"" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" podUID="63fcb9d8-c313-4d0e-9f53-44c3ad272d1d" Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.800867 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" event={"ID":"b0de8493-3f97-4e5e-a28d-5d45f545d645","Type":"ContainerStarted","Data":"9e64ec5a7215a44e96188e7ee3f7ba7a033223793f3f9cf9f607c0b88e060cf6"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.804925 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" event={"ID":"2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282","Type":"ContainerStarted","Data":"9d7db0640c3e0371e8a177c4e69bd5e812448f72deb7ef5bb126bc9f5642475e"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.806030 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" event={"ID":"5c9124b1-4ca0-481f-8474-ea45252bda76","Type":"ContainerStarted","Data":"343d4eee8af91fae5005f6aa99d2cb567609cfa1f73f966511a4318f4afa3099"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.808445 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" event={"ID":"2f4e8661-38a2-44bc-90ae-266f058758a9","Type":"ContainerStarted","Data":"4ca2db2bc11c43c2c8d0c14e99ff8be880272ca90c87f26d97ce655dd5c2c799"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.813028 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.815114 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" event={"ID":"2473d2e8-7c4d-41f5-bd2a-95823fe059f1","Type":"ContainerStarted","Data":"80bdfefb5fc7ef189965f00beda73c0786ec35cfcbe0eb7ad7806c08e425ea5a"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.821116 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" event={"ID":"dcd57b86-afb5-4beb-ae73-830ce75a562f","Type":"ContainerStarted","Data":"86223f5c5fe1e3cec030b633debe90893a22fb4936c8bebc11a55f1840027715"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.822156 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw"] Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.822713 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" podUID="dcd57b86-afb5-4beb-ae73-830ce75a562f" Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.859455 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53989876_5869_4e41_a9c3_c99cea8347bd.slice/crio-13f191c38aa573149e569e1c7cbd6a05e409eaf7e41d8c21fbbfce812e88a0cb WatchSource:0}: Error finding container 13f191c38aa573149e569e1c7cbd6a05e409eaf7e41d8c21fbbfce812e88a0cb: Status 404 returned error can't find the container with id 13f191c38aa573149e569e1c7cbd6a05e409eaf7e41d8c21fbbfce812e88a0cb Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.859617 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" event={"ID":"53318e37-b20a-46eb-8f12-decb087ade91","Type":"ContainerStarted","Data":"9abd8bea346245592f17a128e5b59d3657a2b920017ee6cc5ec5400b5b97bfdc"} Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.862688 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7q4k2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-jz6fw_openstack-operators(53989876-5869-4e41-a9c3-c99cea8347bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.864805 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" podUID="53989876-5869-4e41-a9c3-c99cea8347bd" Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.866957 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" event={"ID":"b1bbb05f-c65b-4ea1-ab19-34b888f49369","Type":"ContainerStarted","Data":"36dda4ea1af9c1b80d73e03d69e128e582a173237677d3a61472214e279278b8"} Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.869944 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-w8w5t"] Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.870203 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" event={"ID":"61877abc-2f6b-40a5-aa19-b8adf5402cc6","Type":"ContainerStarted","Data":"3603e59506c5d7bdcc8ece03da6ef16acf9fa428469ff93c951e75b05b58d82c"} Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.871938 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70a13bde_19ab_4c7c_af9a_15f1fe96d4e1.slice/crio-7a556edf2e2e38df7b32ff417f28d2f26d348a8f3a19993f56834565fc08b71c WatchSource:0}: Error finding container 7a556edf2e2e38df7b32ff417f28d2f26d348a8f3a19993f56834565fc08b71c: Status 404 returned error can't find the container with id 7a556edf2e2e38df7b32ff417f28d2f26d348a8f3a19993f56834565fc08b71c Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.873906 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bc65s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-w8w5t_openstack-operators(70a13bde-19ab-4c7c-af9a-15f1fe96d4e1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.875434 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" podUID="70a13bde-19ab-4c7c-af9a-15f1fe96d4e1" Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.883337 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9"] Feb 19 08:58:11 crc kubenswrapper[4675]: W0219 08:58:11.885272 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a383e65_12c2_43bf_b6b0_f61f20706d7a.slice/crio-bccabf6cdcde63640dafd2b970dce43ff98a4698afecd99eeddde0e04c4c345f WatchSource:0}: Error finding container bccabf6cdcde63640dafd2b970dce43ff98a4698afecd99eeddde0e04c4c345f: Status 404 returned error can't find the container with id bccabf6cdcde63640dafd2b970dce43ff98a4698afecd99eeddde0e04c4c345f Feb 19 08:58:11 crc kubenswrapper[4675]: I0219 08:58:11.886733 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz"] Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.889574 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jghnv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69f8888797-ff4m9_openstack-operators(4fad9fc9-cb88-4aff-89bb-0a2fd609f48a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.891547 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" podUID="4fad9fc9-cb88-4aff-89bb-0a2fd609f48a" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.892327 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-px8qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5db88f68c-x6glz_openstack-operators(1a383e65-12c2-43bf-b6b0-f61f20706d7a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 19 08:58:11 crc kubenswrapper[4675]: E0219 08:58:11.893579 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" podUID="1a383e65-12c2-43bf-b6b0-f61f20706d7a" Feb 19 08:58:12 crc kubenswrapper[4675]: I0219 08:58:12.074424 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:12 crc kubenswrapper[4675]: I0219 08:58:12.074475 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.074547 4675 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.074595 4675 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.074620 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:14.074601637 +0000 UTC m=+875.701691905 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "webhook-server-cert" not found Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.074679 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:14.074658389 +0000 UTC m=+875.701748697 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "metrics-server-cert" not found Feb 19 08:58:12 crc kubenswrapper[4675]: I0219 08:58:12.896498 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" event={"ID":"4fad9fc9-cb88-4aff-89bb-0a2fd609f48a","Type":"ContainerStarted","Data":"b1ef86eee51e1e25386f38ce4ca8b6a0ddbca87fb81646a3014aa91647fd14e6"} Feb 19 08:58:12 crc kubenswrapper[4675]: I0219 08:58:12.898879 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" event={"ID":"b9dc6b16-af6b-470e-a9a7-d59245e8a22e","Type":"ContainerStarted","Data":"b72a5ed4d7b0475d6015bfb2d3ae3718825210a6dd99daaf47914d6e5a4e9f4c"} Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.899491 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" podUID="4fad9fc9-cb88-4aff-89bb-0a2fd609f48a" Feb 19 08:58:12 crc kubenswrapper[4675]: I0219 08:58:12.903496 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" event={"ID":"53989876-5869-4e41-a9c3-c99cea8347bd","Type":"ContainerStarted","Data":"13f191c38aa573149e569e1c7cbd6a05e409eaf7e41d8c21fbbfce812e88a0cb"} Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.905170 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" podUID="53989876-5869-4e41-a9c3-c99cea8347bd" Feb 19 08:58:12 crc kubenswrapper[4675]: I0219 08:58:12.908651 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" event={"ID":"1a383e65-12c2-43bf-b6b0-f61f20706d7a","Type":"ContainerStarted","Data":"bccabf6cdcde63640dafd2b970dce43ff98a4698afecd99eeddde0e04c4c345f"} Feb 19 08:58:12 crc kubenswrapper[4675]: I0219 08:58:12.914283 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" event={"ID":"70a13bde-19ab-4c7c-af9a-15f1fe96d4e1","Type":"ContainerStarted","Data":"7a556edf2e2e38df7b32ff417f28d2f26d348a8f3a19993f56834565fc08b71c"} Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.916184 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" podUID="1a383e65-12c2-43bf-b6b0-f61f20706d7a" Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.919522 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8fb0a33b8d93cf9f84f079af5f2ceb680afada4e44542514959146779f57f64c\\\"\"" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" podUID="63fcb9d8-c313-4d0e-9f53-44c3ad272d1d" Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.920393 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" podUID="dcd57b86-afb5-4beb-ae73-830ce75a562f" Feb 19 08:58:12 crc kubenswrapper[4675]: E0219 08:58:12.920866 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" podUID="70a13bde-19ab-4c7c-af9a-15f1fe96d4e1" Feb 19 08:58:13 crc kubenswrapper[4675]: I0219 08:58:13.601354 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.601512 4675 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.601566 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert podName:a5bfbb7f-0d4a-49d3-bf77-b168ccf33061 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:17.601549915 +0000 UTC m=+879.228640183 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert") pod "infra-operator-controller-manager-79d975b745-cbbkp" (UID: "a5bfbb7f-0d4a-49d3-bf77-b168ccf33061") : secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:13 crc kubenswrapper[4675]: I0219 08:58:13.804392 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.804575 4675 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.804619 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert podName:769e3e7a-a9ea-4d9c-b60b-8643e0c275e1 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:17.804605741 +0000 UTC m=+879.431696009 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" (UID: "769e3e7a-a9ea-4d9c-b60b-8643e0c275e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.926910 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" podUID="53989876-5869-4e41-a9c3-c99cea8347bd" Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.927029 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" podUID="1a383e65-12c2-43bf-b6b0-f61f20706d7a" Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.932117 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" podUID="4fad9fc9-cb88-4aff-89bb-0a2fd609f48a" Feb 19 08:58:13 crc kubenswrapper[4675]: E0219 08:58:13.932160 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" podUID="70a13bde-19ab-4c7c-af9a-15f1fe96d4e1" Feb 19 08:58:14 crc kubenswrapper[4675]: I0219 08:58:14.114480 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:14 crc kubenswrapper[4675]: I0219 08:58:14.114723 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:14 crc kubenswrapper[4675]: E0219 08:58:14.114853 4675 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 19 08:58:14 crc kubenswrapper[4675]: E0219 08:58:14.114903 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:18.11488665 +0000 UTC m=+879.741976918 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "webhook-server-cert" not found Feb 19 08:58:14 crc kubenswrapper[4675]: E0219 08:58:14.115218 4675 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 19 08:58:14 crc kubenswrapper[4675]: E0219 08:58:14.115251 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:18.11524404 +0000 UTC m=+879.742334308 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "metrics-server-cert" not found Feb 19 08:58:17 crc kubenswrapper[4675]: I0219 08:58:17.667262 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:17 crc kubenswrapper[4675]: E0219 08:58:17.667840 4675 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:17 crc kubenswrapper[4675]: E0219 08:58:17.667897 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert podName:a5bfbb7f-0d4a-49d3-bf77-b168ccf33061 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:25.667882375 +0000 UTC m=+887.294972643 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert") pod "infra-operator-controller-manager-79d975b745-cbbkp" (UID: "a5bfbb7f-0d4a-49d3-bf77-b168ccf33061") : secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:17 crc kubenswrapper[4675]: I0219 08:58:17.876319 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:17 crc kubenswrapper[4675]: E0219 08:58:17.876483 4675 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:17 crc kubenswrapper[4675]: E0219 08:58:17.876858 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert podName:769e3e7a-a9ea-4d9c-b60b-8643e0c275e1 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:25.876842261 +0000 UTC m=+887.503932519 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" (UID: "769e3e7a-a9ea-4d9c-b60b-8643e0c275e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:18 crc kubenswrapper[4675]: I0219 08:58:18.180533 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:18 crc kubenswrapper[4675]: I0219 08:58:18.180590 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:18 crc kubenswrapper[4675]: E0219 08:58:18.181984 4675 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 19 08:58:18 crc kubenswrapper[4675]: E0219 08:58:18.182052 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:26.182031252 +0000 UTC m=+887.809121530 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "webhook-server-cert" not found Feb 19 08:58:18 crc kubenswrapper[4675]: E0219 08:58:18.182479 4675 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 19 08:58:18 crc kubenswrapper[4675]: E0219 08:58:18.182507 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:26.182498294 +0000 UTC m=+887.809588562 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "metrics-server-cert" not found Feb 19 08:58:25 crc kubenswrapper[4675]: I0219 08:58:25.688110 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:25 crc kubenswrapper[4675]: E0219 08:58:25.688318 4675 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:25 crc kubenswrapper[4675]: E0219 08:58:25.688902 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert podName:a5bfbb7f-0d4a-49d3-bf77-b168ccf33061 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:41.688885541 +0000 UTC m=+903.315975799 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert") pod "infra-operator-controller-manager-79d975b745-cbbkp" (UID: "a5bfbb7f-0d4a-49d3-bf77-b168ccf33061") : secret "infra-operator-webhook-server-cert" not found Feb 19 08:58:25 crc kubenswrapper[4675]: E0219 08:58:25.759342 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1" Feb 19 08:58:25 crc kubenswrapper[4675]: E0219 08:58:25.759522 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-trbn9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b4d948c87-jhgw4_openstack-operators(4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 08:58:25 crc kubenswrapper[4675]: E0219 08:58:25.760759 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" podUID="4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6" Feb 19 08:58:25 crc kubenswrapper[4675]: I0219 08:58:25.890967 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:25 crc kubenswrapper[4675]: E0219 08:58:25.891259 4675 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:25 crc kubenswrapper[4675]: E0219 08:58:25.891317 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert podName:769e3e7a-a9ea-4d9c-b60b-8643e0c275e1 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:41.891300419 +0000 UTC m=+903.518390687 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" (UID: "769e3e7a-a9ea-4d9c-b60b-8643e0c275e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.036748 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" podUID="4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6" Feb 19 08:58:26 crc kubenswrapper[4675]: I0219 08:58:26.195235 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:26 crc kubenswrapper[4675]: I0219 08:58:26.195285 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.195424 4675 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.195505 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:42.195484512 +0000 UTC m=+903.822574850 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "webhook-server-cert" not found Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.195607 4675 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.195734 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs podName:a4559dd6-5152-40d8-94af-f2fff6c90e84 nodeName:}" failed. No retries permitted until 2026-02-19 08:58:42.195708708 +0000 UTC m=+903.822798986 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs") pod "openstack-operator-controller-manager-7d685cfdd5-zzm7g" (UID: "a4559dd6-5152-40d8-94af-f2fff6c90e84") : secret "metrics-server-cert" not found Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.221906 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da" Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.222120 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-knvkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5b9b8895d5-9zn82_openstack-operators(2473d2e8-7c4d-41f5-bd2a-95823fe059f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.223334 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" podUID="2473d2e8-7c4d-41f5-bd2a-95823fe059f1" Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.824613 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a" Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.824880 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4sd8s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6994f66f48-c25kr_openstack-operators(61877abc-2f6b-40a5-aa19-b8adf5402cc6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 08:58:26 crc kubenswrapper[4675]: E0219 08:58:26.826124 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" podUID="61877abc-2f6b-40a5-aa19-b8adf5402cc6" Feb 19 08:58:27 crc kubenswrapper[4675]: E0219 08:58:27.043017 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" podUID="61877abc-2f6b-40a5-aa19-b8adf5402cc6" Feb 19 08:58:27 crc kubenswrapper[4675]: E0219 08:58:27.045283 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" podUID="2473d2e8-7c4d-41f5-bd2a-95823fe059f1" Feb 19 08:58:27 crc kubenswrapper[4675]: E0219 08:58:27.348828 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Feb 19 08:58:27 crc kubenswrapper[4675]: E0219 08:58:27.349051 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mvndm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-sm2fn_openstack-operators(b9dc6b16-af6b-470e-a9a7-d59245e8a22e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 08:58:27 crc kubenswrapper[4675]: E0219 08:58:27.350763 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" podUID="b9dc6b16-af6b-470e-a9a7-d59245e8a22e" Feb 19 08:58:28 crc kubenswrapper[4675]: E0219 08:58:28.049167 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" podUID="b9dc6b16-af6b-470e-a9a7-d59245e8a22e" Feb 19 08:58:28 crc kubenswrapper[4675]: E0219 08:58:28.061358 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838" Feb 19 08:58:28 crc kubenswrapper[4675]: E0219 08:58:28.061690 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bpqg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-ddhjt_openstack-operators(4e163d5d-3666-4899-acd7-21040c1cc573): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 08:58:28 crc kubenswrapper[4675]: E0219 08:58:28.063060 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" podUID="4e163d5d-3666-4899-acd7-21040c1cc573" Feb 19 08:58:29 crc kubenswrapper[4675]: E0219 08:58:29.057587 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" podUID="4e163d5d-3666-4899-acd7-21040c1cc573" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.089539 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" event={"ID":"5c9124b1-4ca0-481f-8474-ea45252bda76","Type":"ContainerStarted","Data":"2b1c62c47d191048764dda426866cf08c05c483a249df27f6644d37bdf59f5cc"} Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.089602 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.094096 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" event={"ID":"53318e37-b20a-46eb-8f12-decb087ade91","Type":"ContainerStarted","Data":"539d7202bc271c11fc460524b7fe31da296371e1d7103898d2c52ee22f49d19b"} Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.094170 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.095473 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" event={"ID":"78209850-b313-4190-a3a4-674da1afaaba","Type":"ContainerStarted","Data":"41e57ef25d9a3123620f537c8f35b2f18bed347d515e23d1efdf5f2e41d7e2ba"} Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.095818 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.096991 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" event={"ID":"b1bbb05f-c65b-4ea1-ab19-34b888f49369","Type":"ContainerStarted","Data":"7db041cc56453a24f6e7a46f4a35176d85b9f58b3fd38122b3dd232350035a55"} Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.097312 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.100533 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" event={"ID":"2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282","Type":"ContainerStarted","Data":"1f1a95340a1dd87d159c53cb86bc402de0195cbab2357d10a2317f9be146a863"} Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.100732 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.106872 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" podStartSLOduration=4.637623126 podStartE2EDuration="21.106849952s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.565053065 +0000 UTC m=+873.192143323" lastFinishedPulling="2026-02-19 08:58:28.034279891 +0000 UTC m=+889.661370149" observedRunningTime="2026-02-19 08:58:30.103353627 +0000 UTC m=+891.730443895" watchObservedRunningTime="2026-02-19 08:58:30.106849952 +0000 UTC m=+891.733940220" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.131328 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" podStartSLOduration=4.574885412 podStartE2EDuration="21.131300476s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.477828426 +0000 UTC m=+873.104918704" lastFinishedPulling="2026-02-19 08:58:28.0342435 +0000 UTC m=+889.661333768" observedRunningTime="2026-02-19 08:58:30.122682222 +0000 UTC m=+891.749772490" watchObservedRunningTime="2026-02-19 08:58:30.131300476 +0000 UTC m=+891.758390764" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.147959 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" podStartSLOduration=4.002717079 podStartE2EDuration="21.147943298s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.451930123 +0000 UTC m=+873.079020391" lastFinishedPulling="2026-02-19 08:58:28.597156342 +0000 UTC m=+890.224246610" observedRunningTime="2026-02-19 08:58:30.147565748 +0000 UTC m=+891.774656016" watchObservedRunningTime="2026-02-19 08:58:30.147943298 +0000 UTC m=+891.775033566" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.194860 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" podStartSLOduration=4.2275923970000004 podStartE2EDuration="21.194836202s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.067044677 +0000 UTC m=+872.694134945" lastFinishedPulling="2026-02-19 08:58:28.034288482 +0000 UTC m=+889.661378750" observedRunningTime="2026-02-19 08:58:30.191960094 +0000 UTC m=+891.819050362" watchObservedRunningTime="2026-02-19 08:58:30.194836202 +0000 UTC m=+891.821926480" Feb 19 08:58:30 crc kubenswrapper[4675]: I0219 08:58:30.195939 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" podStartSLOduration=3.718451887 podStartE2EDuration="21.195930072s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.641358608 +0000 UTC m=+873.268448866" lastFinishedPulling="2026-02-19 08:58:29.118836783 +0000 UTC m=+890.745927051" observedRunningTime="2026-02-19 08:58:30.168342033 +0000 UTC m=+891.795432301" watchObservedRunningTime="2026-02-19 08:58:30.195930072 +0000 UTC m=+891.823020440" Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.138185 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" event={"ID":"b0de8493-3f97-4e5e-a28d-5d45f545d645","Type":"ContainerStarted","Data":"20c8d63c9cad9642b2941885ca1b7eaf9bc115c93220337f85746f7f6aaffc01"} Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.138545 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.139499 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" event={"ID":"2f4e8661-38a2-44bc-90ae-266f058758a9","Type":"ContainerStarted","Data":"6657801e8777ebd7f61a20c09d6efef96cdc27393bce8ca71777c599d8dbc246"} Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.140097 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.142621 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" event={"ID":"f79e0726-e5c1-4536-b06b-40bab3849bb0","Type":"ContainerStarted","Data":"9074e3d9042fadf13335d9419d6b07a86fd6942a30a557abc9189ae40d97ca6c"} Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.142665 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.155457 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" podStartSLOduration=6.605907323 podStartE2EDuration="23.1554396s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.483625964 +0000 UTC m=+873.110716232" lastFinishedPulling="2026-02-19 08:58:28.033158241 +0000 UTC m=+889.660248509" observedRunningTime="2026-02-19 08:58:32.153055645 +0000 UTC m=+893.780145913" watchObservedRunningTime="2026-02-19 08:58:32.1554396 +0000 UTC m=+893.782529868" Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.172193 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" podStartSLOduration=5.657252832 podStartE2EDuration="23.172170664s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.082138127 +0000 UTC m=+872.709228395" lastFinishedPulling="2026-02-19 08:58:28.597055959 +0000 UTC m=+890.224146227" observedRunningTime="2026-02-19 08:58:32.170141259 +0000 UTC m=+893.797231527" watchObservedRunningTime="2026-02-19 08:58:32.172170664 +0000 UTC m=+893.799260932" Feb 19 08:58:32 crc kubenswrapper[4675]: I0219 08:58:32.187911 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" podStartSLOduration=6.726402795 podStartE2EDuration="23.187893511s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.572798946 +0000 UTC m=+873.199889214" lastFinishedPulling="2026-02-19 08:58:28.034289662 +0000 UTC m=+889.661379930" observedRunningTime="2026-02-19 08:58:32.186922135 +0000 UTC m=+893.814012413" watchObservedRunningTime="2026-02-19 08:58:32.187893511 +0000 UTC m=+893.814983779" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.155567 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" event={"ID":"53989876-5869-4e41-a9c3-c99cea8347bd","Type":"ContainerStarted","Data":"3004cc964a49adb74dcbaadc4729b76046fc882e60e57880abd145f7596aed04"} Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.156157 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.158073 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" event={"ID":"1a383e65-12c2-43bf-b6b0-f61f20706d7a","Type":"ContainerStarted","Data":"94defd0825bdc831df7fbb849065e2c117628bd226019318c653f7b984d7309c"} Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.158280 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.159260 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" event={"ID":"70a13bde-19ab-4c7c-af9a-15f1fe96d4e1","Type":"ContainerStarted","Data":"e910aa4fec64ce59f2b888dbd6d934689008f6969408c54fee0d3dbb0112fdba"} Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.159441 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.160238 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" event={"ID":"dcd57b86-afb5-4beb-ae73-830ce75a562f","Type":"ContainerStarted","Data":"9e4baf995bf0e461368d9362af15b50c26cf8d2ba2be3952a8f65cdb73b1bcbf"} Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.160392 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.161090 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" event={"ID":"63fcb9d8-c313-4d0e-9f53-44c3ad272d1d","Type":"ContainerStarted","Data":"a51c9d9fc13d1c7f13ce83633fc0a8990602c27d35d35e8a5aa1a99ac37db153"} Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.161715 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.162623 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" event={"ID":"4fad9fc9-cb88-4aff-89bb-0a2fd609f48a","Type":"ContainerStarted","Data":"9b5380e4accd027809a07213cf5ed33a95daab8731423d278a18ddabdb6b3c0a"} Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.163034 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.164571 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" event={"ID":"7f7fd1c2-d277-402d-a762-0ebfddc48226","Type":"ContainerStarted","Data":"b211e982d9d755581ce9b362dee782256370337dc9852625f94fa5399e2ed02d"} Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.164597 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.172816 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" podStartSLOduration=7.231146466 podStartE2EDuration="25.17279962s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.862487215 +0000 UTC m=+873.489577483" lastFinishedPulling="2026-02-19 08:58:29.804140369 +0000 UTC m=+891.431230637" observedRunningTime="2026-02-19 08:58:34.170933979 +0000 UTC m=+895.798024247" watchObservedRunningTime="2026-02-19 08:58:34.17279962 +0000 UTC m=+895.799889888" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.270179 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" podStartSLOduration=2.544994114 podStartE2EDuration="24.270146684s" podCreationTimestamp="2026-02-19 08:58:10 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.892179472 +0000 UTC m=+873.519269740" lastFinishedPulling="2026-02-19 08:58:33.617332042 +0000 UTC m=+895.244422310" observedRunningTime="2026-02-19 08:58:34.201606933 +0000 UTC m=+895.828697211" watchObservedRunningTime="2026-02-19 08:58:34.270146684 +0000 UTC m=+895.897236952" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.271180 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" podStartSLOduration=7.012219909 podStartE2EDuration="25.271172752s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.474794164 +0000 UTC m=+873.101884432" lastFinishedPulling="2026-02-19 08:58:29.733747007 +0000 UTC m=+891.360837275" observedRunningTime="2026-02-19 08:58:34.265050706 +0000 UTC m=+895.892140984" watchObservedRunningTime="2026-02-19 08:58:34.271172752 +0000 UTC m=+895.898263040" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.342695 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" podStartSLOduration=6.480444539 podStartE2EDuration="24.342678705s" podCreationTimestamp="2026-02-19 08:58:10 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.873505935 +0000 UTC m=+873.500596203" lastFinishedPulling="2026-02-19 08:58:29.735740101 +0000 UTC m=+891.362830369" observedRunningTime="2026-02-19 08:58:34.336118827 +0000 UTC m=+895.963209095" watchObservedRunningTime="2026-02-19 08:58:34.342678705 +0000 UTC m=+895.969768973" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.343206 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" podStartSLOduration=7.11535101 podStartE2EDuration="25.343201069s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.648902953 +0000 UTC m=+873.275993211" lastFinishedPulling="2026-02-19 08:58:29.876753002 +0000 UTC m=+891.503843270" observedRunningTime="2026-02-19 08:58:34.304344013 +0000 UTC m=+895.931434281" watchObservedRunningTime="2026-02-19 08:58:34.343201069 +0000 UTC m=+895.970291337" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.357683 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" podStartSLOduration=3.397150778 podStartE2EDuration="25.357663812s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.675485335 +0000 UTC m=+873.302575603" lastFinishedPulling="2026-02-19 08:58:33.635998369 +0000 UTC m=+895.263088637" observedRunningTime="2026-02-19 08:58:34.352335447 +0000 UTC m=+895.979425725" watchObservedRunningTime="2026-02-19 08:58:34.357663812 +0000 UTC m=+895.984754080" Feb 19 08:58:34 crc kubenswrapper[4675]: I0219 08:58:34.365547 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" podStartSLOduration=3.618365318 podStartE2EDuration="25.365532436s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.889377916 +0000 UTC m=+873.516468184" lastFinishedPulling="2026-02-19 08:58:33.636545034 +0000 UTC m=+895.263635302" observedRunningTime="2026-02-19 08:58:34.364515268 +0000 UTC m=+895.991605536" watchObservedRunningTime="2026-02-19 08:58:34.365532436 +0000 UTC m=+895.992622704" Feb 19 08:58:39 crc kubenswrapper[4675]: I0219 08:58:39.915185 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-bj7kj" Feb 19 08:58:39 crc kubenswrapper[4675]: I0219 08:58:39.925738 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-c775s" Feb 19 08:58:39 crc kubenswrapper[4675]: I0219 08:58:39.961813 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-glh49" Feb 19 08:58:39 crc kubenswrapper[4675]: I0219 08:58:39.981439 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-v7lcl" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.013601 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-bwwwp" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.092307 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-z2wq2" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.343603 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-kqz2s" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.359995 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-s25g8" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.367131 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-8k6xs" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.411096 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-w8w5t" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.422881 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-ff4m9" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.521202 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-r8bkh" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.531531 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dr2nj" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.596185 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-jz6fw" Feb 19 08:58:40 crc kubenswrapper[4675]: I0219 08:58:40.750618 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x6glz" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.230094 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" event={"ID":"4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6","Type":"ContainerStarted","Data":"7adb1ea3f7809d0e078b35076d0c3db8350044c830e132321340795cdb53df46"} Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.231043 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.249368 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" podStartSLOduration=3.099887055 podStartE2EDuration="32.24934992s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.444712027 +0000 UTC m=+873.071802295" lastFinishedPulling="2026-02-19 08:58:40.594174882 +0000 UTC m=+902.221265160" observedRunningTime="2026-02-19 08:58:41.243572713 +0000 UTC m=+902.870663001" watchObservedRunningTime="2026-02-19 08:58:41.24934992 +0000 UTC m=+902.876440188" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.734488 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.752381 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5bfbb7f-0d4a-49d3-bf77-b168ccf33061-cert\") pod \"infra-operator-controller-manager-79d975b745-cbbkp\" (UID: \"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.869339 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-twvvd" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.878478 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.937139 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.943280 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/769e3e7a-a9ea-4d9c-b60b-8643e0c275e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8\" (UID: \"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.984815 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-h9pmt" Feb 19 08:58:41 crc kubenswrapper[4675]: I0219 08:58:41.992934 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.199346 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp"] Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.236978 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" event={"ID":"2473d2e8-7c4d-41f5-bd2a-95823fe059f1","Type":"ContainerStarted","Data":"2cfd18e1a135c2ad32d843a021f7b2a522cd9431b16cb2b38ac91fe0a7ae4af6"} Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.237997 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.240886 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.240917 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.241828 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" event={"ID":"61877abc-2f6b-40a5-aa19-b8adf5402cc6","Type":"ContainerStarted","Data":"2437990ac6698323d3cf95b472d607c6147555efa7e9e44001bf88d68a066c8b"} Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.242112 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.244204 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" event={"ID":"b9dc6b16-af6b-470e-a9a7-d59245e8a22e","Type":"ContainerStarted","Data":"76afb52ae783c67341fe8c870db881ed569388f41618a67d5a2c433fa391e87c"} Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.245837 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-metrics-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.249711 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" event={"ID":"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061","Type":"ContainerStarted","Data":"4e111ab516b78bcadfbefcfba38267241542547bd8e71e35e788e62d57e34b8d"} Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.251216 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4559dd6-5152-40d8-94af-f2fff6c90e84-webhook-certs\") pod \"openstack-operator-controller-manager-7d685cfdd5-zzm7g\" (UID: \"a4559dd6-5152-40d8-94af-f2fff6c90e84\") " pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.257308 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" podStartSLOduration=3.162921886 podStartE2EDuration="33.257283489s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.449953969 +0000 UTC m=+873.077044257" lastFinishedPulling="2026-02-19 08:58:41.544315592 +0000 UTC m=+903.171405860" observedRunningTime="2026-02-19 08:58:42.25400796 +0000 UTC m=+903.881098238" watchObservedRunningTime="2026-02-19 08:58:42.257283489 +0000 UTC m=+903.884373757" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.280228 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" podStartSLOduration=3.342543156 podStartE2EDuration="33.280204092s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.605829473 +0000 UTC m=+873.232919741" lastFinishedPulling="2026-02-19 08:58:41.543490409 +0000 UTC m=+903.170580677" observedRunningTime="2026-02-19 08:58:42.272604976 +0000 UTC m=+903.899695244" watchObservedRunningTime="2026-02-19 08:58:42.280204092 +0000 UTC m=+903.907294360" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.280524 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-jwl72" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.290071 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.300377 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-sm2fn" podStartSLOduration=2.624767832 podStartE2EDuration="32.300357519s" podCreationTimestamp="2026-02-19 08:58:10 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.868356375 +0000 UTC m=+873.495446643" lastFinishedPulling="2026-02-19 08:58:41.543946042 +0000 UTC m=+903.171036330" observedRunningTime="2026-02-19 08:58:42.293710519 +0000 UTC m=+903.920800787" watchObservedRunningTime="2026-02-19 08:58:42.300357519 +0000 UTC m=+903.927447787" Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.359932 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8"] Feb 19 08:58:42 crc kubenswrapper[4675]: W0219 08:58:42.363418 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod769e3e7a_a9ea_4d9c_b60b_8643e0c275e1.slice/crio-877ea1ddca7ed5bb3eb2f14f5e3e894e37bb7d9792025e782d2e08de07c3d33d WatchSource:0}: Error finding container 877ea1ddca7ed5bb3eb2f14f5e3e894e37bb7d9792025e782d2e08de07c3d33d: Status 404 returned error can't find the container with id 877ea1ddca7ed5bb3eb2f14f5e3e894e37bb7d9792025e782d2e08de07c3d33d Feb 19 08:58:42 crc kubenswrapper[4675]: I0219 08:58:42.770300 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g"] Feb 19 08:58:43 crc kubenswrapper[4675]: I0219 08:58:43.257667 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" event={"ID":"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1","Type":"ContainerStarted","Data":"877ea1ddca7ed5bb3eb2f14f5e3e894e37bb7d9792025e782d2e08de07c3d33d"} Feb 19 08:58:43 crc kubenswrapper[4675]: I0219 08:58:43.261093 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" event={"ID":"a4559dd6-5152-40d8-94af-f2fff6c90e84","Type":"ContainerStarted","Data":"384bfff93db4ce2eab59da963eda6f873237de027a4929f52d6a6f9f6391e84a"} Feb 19 08:58:43 crc kubenswrapper[4675]: I0219 08:58:43.261174 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" event={"ID":"a4559dd6-5152-40d8-94af-f2fff6c90e84","Type":"ContainerStarted","Data":"920550e7dc3d12e721e5d3d2bc5d5b54531f06357bf82e348acfbe1190fd9bb5"} Feb 19 08:58:43 crc kubenswrapper[4675]: I0219 08:58:43.261658 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.185925 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" podStartSLOduration=34.185901429 podStartE2EDuration="34.185901429s" podCreationTimestamp="2026-02-19 08:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 08:58:43.293049566 +0000 UTC m=+904.920139834" watchObservedRunningTime="2026-02-19 08:58:44.185901429 +0000 UTC m=+905.812991687" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.198978 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pstjt"] Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.204763 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.211373 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pstjt"] Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.267244 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" event={"ID":"4e163d5d-3666-4899-acd7-21040c1cc573","Type":"ContainerStarted","Data":"cbee512999659b5335cb14e4c67f1689728d8b86b96d90aceff9c971395a28bb"} Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.267455 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.277891 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-catalog-content\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.278003 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-utilities\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.278026 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2grmv\" (UniqueName: \"kubernetes.io/projected/61cd54b1-635f-4a71-9e83-a7dba862481c-kube-api-access-2grmv\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.284083 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" podStartSLOduration=3.428501841 podStartE2EDuration="35.284061166s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:11.669499503 +0000 UTC m=+873.296589771" lastFinishedPulling="2026-02-19 08:58:43.525058828 +0000 UTC m=+905.152149096" observedRunningTime="2026-02-19 08:58:44.280938541 +0000 UTC m=+905.908028809" watchObservedRunningTime="2026-02-19 08:58:44.284061166 +0000 UTC m=+905.911151434" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.378963 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-catalog-content\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.379089 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-utilities\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.379114 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2grmv\" (UniqueName: \"kubernetes.io/projected/61cd54b1-635f-4a71-9e83-a7dba862481c-kube-api-access-2grmv\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.381008 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-catalog-content\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.381886 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-utilities\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.405675 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2grmv\" (UniqueName: \"kubernetes.io/projected/61cd54b1-635f-4a71-9e83-a7dba862481c-kube-api-access-2grmv\") pod \"community-operators-pstjt\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:44 crc kubenswrapper[4675]: I0219 08:58:44.533198 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.765085 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tgnzv"] Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.768963 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.776391 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tgnzv"] Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.837259 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-catalog-content\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.837339 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf55r\" (UniqueName: \"kubernetes.io/projected/b0ae0508-cd23-4122-bbe1-44686f753318-kube-api-access-cf55r\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.837515 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-utilities\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.938547 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf55r\" (UniqueName: \"kubernetes.io/projected/b0ae0508-cd23-4122-bbe1-44686f753318-kube-api-access-cf55r\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.938624 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-utilities\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.938720 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-catalog-content\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.939116 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-utilities\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.939169 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-catalog-content\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:48 crc kubenswrapper[4675]: I0219 08:58:48.960569 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf55r\" (UniqueName: \"kubernetes.io/projected/b0ae0508-cd23-4122-bbe1-44686f753318-kube-api-access-cf55r\") pod \"certified-operators-tgnzv\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:49 crc kubenswrapper[4675]: I0219 08:58:49.108083 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:58:50 crc kubenswrapper[4675]: I0219 08:58:50.022168 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-9zn82" Feb 19 08:58:50 crc kubenswrapper[4675]: I0219 08:58:50.106153 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-jhgw4" Feb 19 08:58:50 crc kubenswrapper[4675]: I0219 08:58:50.213729 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c25kr" Feb 19 08:58:50 crc kubenswrapper[4675]: I0219 08:58:50.431408 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-ddhjt" Feb 19 08:58:52 crc kubenswrapper[4675]: I0219 08:58:52.295449 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7d685cfdd5-zzm7g" Feb 19 08:58:54 crc kubenswrapper[4675]: I0219 08:58:54.453806 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pstjt"] Feb 19 08:58:54 crc kubenswrapper[4675]: I0219 08:58:54.705204 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tgnzv"] Feb 19 08:58:54 crc kubenswrapper[4675]: W0219 08:58:54.711441 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0ae0508_cd23_4122_bbe1_44686f753318.slice/crio-971ea3179b7ce4fb6dd344ff764a8609bb2756aa003176506866135451ec6730 WatchSource:0}: Error finding container 971ea3179b7ce4fb6dd344ff764a8609bb2756aa003176506866135451ec6730: Status 404 returned error can't find the container with id 971ea3179b7ce4fb6dd344ff764a8609bb2756aa003176506866135451ec6730 Feb 19 08:58:55 crc kubenswrapper[4675]: E0219 08:58:55.124509 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e6f7c2a75883f63d270378b283faeee4c4c14fbd74b509c7da82621166f07b24" Feb 19 08:58:55 crc kubenswrapper[4675]: E0219 08:58:55.126346 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e6f7c2a75883f63d270378b283faeee4c4c14fbd74b509c7da82621166f07b24,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4kgrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8_openstack-operators(769e3e7a-a9ea-4d9c-b60b-8643e0c275e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 08:58:55 crc kubenswrapper[4675]: E0219 08:58:55.128258 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" podUID="769e3e7a-a9ea-4d9c-b60b-8643e0c275e1" Feb 19 08:58:55 crc kubenswrapper[4675]: I0219 08:58:55.358453 4675 generic.go:334] "Generic (PLEG): container finished" podID="b0ae0508-cd23-4122-bbe1-44686f753318" containerID="9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b" exitCode=0 Feb 19 08:58:55 crc kubenswrapper[4675]: I0219 08:58:55.358512 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgnzv" event={"ID":"b0ae0508-cd23-4122-bbe1-44686f753318","Type":"ContainerDied","Data":"9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b"} Feb 19 08:58:55 crc kubenswrapper[4675]: I0219 08:58:55.358909 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgnzv" event={"ID":"b0ae0508-cd23-4122-bbe1-44686f753318","Type":"ContainerStarted","Data":"971ea3179b7ce4fb6dd344ff764a8609bb2756aa003176506866135451ec6730"} Feb 19 08:58:55 crc kubenswrapper[4675]: I0219 08:58:55.369052 4675 generic.go:334] "Generic (PLEG): container finished" podID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerID="95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2" exitCode=0 Feb 19 08:58:55 crc kubenswrapper[4675]: I0219 08:58:55.369484 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pstjt" event={"ID":"61cd54b1-635f-4a71-9e83-a7dba862481c","Type":"ContainerDied","Data":"95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2"} Feb 19 08:58:55 crc kubenswrapper[4675]: I0219 08:58:55.369542 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pstjt" event={"ID":"61cd54b1-635f-4a71-9e83-a7dba862481c","Type":"ContainerStarted","Data":"b5da853a6d959978c3bc9e2a6ddc757d1f3a41a7c4d049996918f300db7809ce"} Feb 19 08:58:55 crc kubenswrapper[4675]: E0219 08:58:55.371233 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e6f7c2a75883f63d270378b283faeee4c4c14fbd74b509c7da82621166f07b24\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" podUID="769e3e7a-a9ea-4d9c-b60b-8643e0c275e1" Feb 19 08:58:56 crc kubenswrapper[4675]: I0219 08:58:56.391200 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" event={"ID":"a5bfbb7f-0d4a-49d3-bf77-b168ccf33061","Type":"ContainerStarted","Data":"965a84fdd25b7aafa261713a7df56451497a14fb5e77686b0c0cd5282ef394de"} Feb 19 08:58:56 crc kubenswrapper[4675]: I0219 08:58:56.392069 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:58:56 crc kubenswrapper[4675]: I0219 08:58:56.414439 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" podStartSLOduration=34.488787218 podStartE2EDuration="47.414416273s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:42.21681378 +0000 UTC m=+903.843904048" lastFinishedPulling="2026-02-19 08:58:55.142442835 +0000 UTC m=+916.769533103" observedRunningTime="2026-02-19 08:58:56.409000454 +0000 UTC m=+918.036090722" watchObservedRunningTime="2026-02-19 08:58:56.414416273 +0000 UTC m=+918.041506531" Feb 19 08:58:57 crc kubenswrapper[4675]: I0219 08:58:57.401663 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgnzv" event={"ID":"b0ae0508-cd23-4122-bbe1-44686f753318","Type":"ContainerStarted","Data":"ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff"} Feb 19 08:58:57 crc kubenswrapper[4675]: I0219 08:58:57.405200 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pstjt" event={"ID":"61cd54b1-635f-4a71-9e83-a7dba862481c","Type":"ContainerStarted","Data":"499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2"} Feb 19 08:58:58 crc kubenswrapper[4675]: I0219 08:58:58.414600 4675 generic.go:334] "Generic (PLEG): container finished" podID="b0ae0508-cd23-4122-bbe1-44686f753318" containerID="ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff" exitCode=0 Feb 19 08:58:58 crc kubenswrapper[4675]: I0219 08:58:58.414682 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgnzv" event={"ID":"b0ae0508-cd23-4122-bbe1-44686f753318","Type":"ContainerDied","Data":"ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff"} Feb 19 08:58:58 crc kubenswrapper[4675]: I0219 08:58:58.418176 4675 generic.go:334] "Generic (PLEG): container finished" podID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerID="499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2" exitCode=0 Feb 19 08:58:58 crc kubenswrapper[4675]: I0219 08:58:58.418213 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pstjt" event={"ID":"61cd54b1-635f-4a71-9e83-a7dba862481c","Type":"ContainerDied","Data":"499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2"} Feb 19 08:59:00 crc kubenswrapper[4675]: I0219 08:59:00.440085 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pstjt" event={"ID":"61cd54b1-635f-4a71-9e83-a7dba862481c","Type":"ContainerStarted","Data":"cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34"} Feb 19 08:59:00 crc kubenswrapper[4675]: I0219 08:59:00.442512 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgnzv" event={"ID":"b0ae0508-cd23-4122-bbe1-44686f753318","Type":"ContainerStarted","Data":"73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a"} Feb 19 08:59:00 crc kubenswrapper[4675]: I0219 08:59:00.456927 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pstjt" podStartSLOduration=12.589000467 podStartE2EDuration="16.456911576s" podCreationTimestamp="2026-02-19 08:58:44 +0000 UTC" firstStartedPulling="2026-02-19 08:58:55.370802681 +0000 UTC m=+916.997892949" lastFinishedPulling="2026-02-19 08:58:59.23871379 +0000 UTC m=+920.865804058" observedRunningTime="2026-02-19 08:59:00.456111894 +0000 UTC m=+922.083202162" watchObservedRunningTime="2026-02-19 08:59:00.456911576 +0000 UTC m=+922.084001844" Feb 19 08:59:00 crc kubenswrapper[4675]: I0219 08:59:00.480878 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tgnzv" podStartSLOduration=8.143876512 podStartE2EDuration="12.480858602s" podCreationTimestamp="2026-02-19 08:58:48 +0000 UTC" firstStartedPulling="2026-02-19 08:58:55.360332185 +0000 UTC m=+916.987422453" lastFinishedPulling="2026-02-19 08:58:59.697314275 +0000 UTC m=+921.324404543" observedRunningTime="2026-02-19 08:59:00.476502012 +0000 UTC m=+922.103592280" watchObservedRunningTime="2026-02-19 08:59:00.480858602 +0000 UTC m=+922.107948870" Feb 19 08:59:01 crc kubenswrapper[4675]: I0219 08:59:01.885417 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-cbbkp" Feb 19 08:59:04 crc kubenswrapper[4675]: I0219 08:59:04.533413 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:59:04 crc kubenswrapper[4675]: I0219 08:59:04.533566 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:59:04 crc kubenswrapper[4675]: I0219 08:59:04.601348 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:59:05 crc kubenswrapper[4675]: I0219 08:59:05.519527 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:59:05 crc kubenswrapper[4675]: I0219 08:59:05.568856 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pstjt"] Feb 19 08:59:07 crc kubenswrapper[4675]: I0219 08:59:07.485883 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pstjt" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="registry-server" containerID="cri-o://cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34" gracePeriod=2 Feb 19 08:59:07 crc kubenswrapper[4675]: I0219 08:59:07.884632 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.040166 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-catalog-content\") pod \"61cd54b1-635f-4a71-9e83-a7dba862481c\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.040290 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2grmv\" (UniqueName: \"kubernetes.io/projected/61cd54b1-635f-4a71-9e83-a7dba862481c-kube-api-access-2grmv\") pod \"61cd54b1-635f-4a71-9e83-a7dba862481c\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.040349 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-utilities\") pod \"61cd54b1-635f-4a71-9e83-a7dba862481c\" (UID: \"61cd54b1-635f-4a71-9e83-a7dba862481c\") " Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.041792 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-utilities" (OuterVolumeSpecName: "utilities") pod "61cd54b1-635f-4a71-9e83-a7dba862481c" (UID: "61cd54b1-635f-4a71-9e83-a7dba862481c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.063103 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61cd54b1-635f-4a71-9e83-a7dba862481c-kube-api-access-2grmv" (OuterVolumeSpecName: "kube-api-access-2grmv") pod "61cd54b1-635f-4a71-9e83-a7dba862481c" (UID: "61cd54b1-635f-4a71-9e83-a7dba862481c"). InnerVolumeSpecName "kube-api-access-2grmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.114323 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61cd54b1-635f-4a71-9e83-a7dba862481c" (UID: "61cd54b1-635f-4a71-9e83-a7dba862481c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.142363 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.142394 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2grmv\" (UniqueName: \"kubernetes.io/projected/61cd54b1-635f-4a71-9e83-a7dba862481c-kube-api-access-2grmv\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.142407 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cd54b1-635f-4a71-9e83-a7dba862481c-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.494179 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" event={"ID":"769e3e7a-a9ea-4d9c-b60b-8643e0c275e1","Type":"ContainerStarted","Data":"9a361cba0178973cb0be6ea94a5779a7cdeebc3f8c1b7cac488da6cdc2ef4071"} Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.494981 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.496488 4675 generic.go:334] "Generic (PLEG): container finished" podID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerID="cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34" exitCode=0 Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.496525 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pstjt" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.496535 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pstjt" event={"ID":"61cd54b1-635f-4a71-9e83-a7dba862481c","Type":"ContainerDied","Data":"cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34"} Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.496572 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pstjt" event={"ID":"61cd54b1-635f-4a71-9e83-a7dba862481c","Type":"ContainerDied","Data":"b5da853a6d959978c3bc9e2a6ddc757d1f3a41a7c4d049996918f300db7809ce"} Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.496592 4675 scope.go:117] "RemoveContainer" containerID="cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.513452 4675 scope.go:117] "RemoveContainer" containerID="499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.519660 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" podStartSLOduration=34.224982387 podStartE2EDuration="59.51961838s" podCreationTimestamp="2026-02-19 08:58:09 +0000 UTC" firstStartedPulling="2026-02-19 08:58:42.365568121 +0000 UTC m=+903.992658389" lastFinishedPulling="2026-02-19 08:59:07.660204114 +0000 UTC m=+929.287294382" observedRunningTime="2026-02-19 08:59:08.516300029 +0000 UTC m=+930.143390297" watchObservedRunningTime="2026-02-19 08:59:08.51961838 +0000 UTC m=+930.146708648" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.535745 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pstjt"] Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.542675 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pstjt"] Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.550358 4675 scope.go:117] "RemoveContainer" containerID="95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.566711 4675 scope.go:117] "RemoveContainer" containerID="cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34" Feb 19 08:59:08 crc kubenswrapper[4675]: E0219 08:59:08.567174 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34\": container with ID starting with cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34 not found: ID does not exist" containerID="cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.567218 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34"} err="failed to get container status \"cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34\": rpc error: code = NotFound desc = could not find container \"cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34\": container with ID starting with cf5b28ef02abb63bc54d7602437c9061bfdc2ad02a78f928adbc9c22f63bdf34 not found: ID does not exist" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.567248 4675 scope.go:117] "RemoveContainer" containerID="499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2" Feb 19 08:59:08 crc kubenswrapper[4675]: E0219 08:59:08.567563 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2\": container with ID starting with 499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2 not found: ID does not exist" containerID="499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.567604 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2"} err="failed to get container status \"499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2\": rpc error: code = NotFound desc = could not find container \"499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2\": container with ID starting with 499a257b1e2dc09f143e79f07654cd7d1e420d96b6a02b3406b99446e5ec94e2 not found: ID does not exist" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.567631 4675 scope.go:117] "RemoveContainer" containerID="95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2" Feb 19 08:59:08 crc kubenswrapper[4675]: E0219 08:59:08.567999 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2\": container with ID starting with 95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2 not found: ID does not exist" containerID="95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2" Feb 19 08:59:08 crc kubenswrapper[4675]: I0219 08:59:08.568026 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2"} err="failed to get container status \"95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2\": rpc error: code = NotFound desc = could not find container \"95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2\": container with ID starting with 95a2b354d1b66bf523029ea4fe1394137ee1f981c942a10405d7dc4cb67883f2 not found: ID does not exist" Feb 19 08:59:09 crc kubenswrapper[4675]: I0219 08:59:09.116410 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" path="/var/lib/kubelet/pods/61cd54b1-635f-4a71-9e83-a7dba862481c/volumes" Feb 19 08:59:09 crc kubenswrapper[4675]: I0219 08:59:09.118600 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:59:09 crc kubenswrapper[4675]: I0219 08:59:09.118694 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:59:09 crc kubenswrapper[4675]: I0219 08:59:09.164833 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:59:09 crc kubenswrapper[4675]: I0219 08:59:09.554496 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.254202 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sv"] Feb 19 08:59:11 crc kubenswrapper[4675]: E0219 08:59:11.254971 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="extract-content" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.254987 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="extract-content" Feb 19 08:59:11 crc kubenswrapper[4675]: E0219 08:59:11.255007 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="registry-server" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.255015 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="registry-server" Feb 19 08:59:11 crc kubenswrapper[4675]: E0219 08:59:11.255028 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="extract-utilities" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.255033 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="extract-utilities" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.255173 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="61cd54b1-635f-4a71-9e83-a7dba862481c" containerName="registry-server" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.256127 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.260972 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sv"] Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.389682 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2bsl\" (UniqueName: \"kubernetes.io/projected/c5f740aa-7df7-4bb0-a388-1570876749df-kube-api-access-r2bsl\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.389779 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-catalog-content\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.389875 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-utilities\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.431885 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tgnzv"] Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.491175 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2bsl\" (UniqueName: \"kubernetes.io/projected/c5f740aa-7df7-4bb0-a388-1570876749df-kube-api-access-r2bsl\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.491255 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-catalog-content\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.491325 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-utilities\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.491629 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-catalog-content\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.491707 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-utilities\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.512440 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2bsl\" (UniqueName: \"kubernetes.io/projected/c5f740aa-7df7-4bb0-a388-1570876749df-kube-api-access-r2bsl\") pod \"redhat-marketplace-6z8sv\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.515440 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tgnzv" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="registry-server" containerID="cri-o://73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a" gracePeriod=2 Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.580526 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:11 crc kubenswrapper[4675]: I0219 08:59:11.993302 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:59:12 crc kubenswrapper[4675]: W0219 08:59:12.038664 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5f740aa_7df7_4bb0_a388_1570876749df.slice/crio-16da5ec17fb32efeed950f4489fc46a30587b5ea74723f3828cdbad4d90680a0 WatchSource:0}: Error finding container 16da5ec17fb32efeed950f4489fc46a30587b5ea74723f3828cdbad4d90680a0: Status 404 returned error can't find the container with id 16da5ec17fb32efeed950f4489fc46a30587b5ea74723f3828cdbad4d90680a0 Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.039792 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sv"] Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.102329 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf55r\" (UniqueName: \"kubernetes.io/projected/b0ae0508-cd23-4122-bbe1-44686f753318-kube-api-access-cf55r\") pod \"b0ae0508-cd23-4122-bbe1-44686f753318\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.102395 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-utilities\") pod \"b0ae0508-cd23-4122-bbe1-44686f753318\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.102465 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-catalog-content\") pod \"b0ae0508-cd23-4122-bbe1-44686f753318\" (UID: \"b0ae0508-cd23-4122-bbe1-44686f753318\") " Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.103122 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-utilities" (OuterVolumeSpecName: "utilities") pod "b0ae0508-cd23-4122-bbe1-44686f753318" (UID: "b0ae0508-cd23-4122-bbe1-44686f753318"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.108476 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ae0508-cd23-4122-bbe1-44686f753318-kube-api-access-cf55r" (OuterVolumeSpecName: "kube-api-access-cf55r") pod "b0ae0508-cd23-4122-bbe1-44686f753318" (UID: "b0ae0508-cd23-4122-bbe1-44686f753318"). InnerVolumeSpecName "kube-api-access-cf55r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.148840 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0ae0508-cd23-4122-bbe1-44686f753318" (UID: "b0ae0508-cd23-4122-bbe1-44686f753318"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.203387 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.203419 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf55r\" (UniqueName: \"kubernetes.io/projected/b0ae0508-cd23-4122-bbe1-44686f753318-kube-api-access-cf55r\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.203431 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae0508-cd23-4122-bbe1-44686f753318-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.526561 4675 generic.go:334] "Generic (PLEG): container finished" podID="b0ae0508-cd23-4122-bbe1-44686f753318" containerID="73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a" exitCode=0 Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.526603 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgnzv" event={"ID":"b0ae0508-cd23-4122-bbe1-44686f753318","Type":"ContainerDied","Data":"73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a"} Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.526628 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgnzv" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.526708 4675 scope.go:117] "RemoveContainer" containerID="73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.526693 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgnzv" event={"ID":"b0ae0508-cd23-4122-bbe1-44686f753318","Type":"ContainerDied","Data":"971ea3179b7ce4fb6dd344ff764a8609bb2756aa003176506866135451ec6730"} Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.528389 4675 generic.go:334] "Generic (PLEG): container finished" podID="c5f740aa-7df7-4bb0-a388-1570876749df" containerID="8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a" exitCode=0 Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.528422 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sv" event={"ID":"c5f740aa-7df7-4bb0-a388-1570876749df","Type":"ContainerDied","Data":"8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a"} Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.528444 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sv" event={"ID":"c5f740aa-7df7-4bb0-a388-1570876749df","Type":"ContainerStarted","Data":"16da5ec17fb32efeed950f4489fc46a30587b5ea74723f3828cdbad4d90680a0"} Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.556247 4675 scope.go:117] "RemoveContainer" containerID="ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.577100 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tgnzv"] Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.585498 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tgnzv"] Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.595608 4675 scope.go:117] "RemoveContainer" containerID="9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.610168 4675 scope.go:117] "RemoveContainer" containerID="73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a" Feb 19 08:59:12 crc kubenswrapper[4675]: E0219 08:59:12.610623 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a\": container with ID starting with 73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a not found: ID does not exist" containerID="73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.610681 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a"} err="failed to get container status \"73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a\": rpc error: code = NotFound desc = could not find container \"73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a\": container with ID starting with 73d13f4d6c2d6e63ad833034b63151f03773556a3b9054a2b0f868029c7e543a not found: ID does not exist" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.610717 4675 scope.go:117] "RemoveContainer" containerID="ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff" Feb 19 08:59:12 crc kubenswrapper[4675]: E0219 08:59:12.611001 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff\": container with ID starting with ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff not found: ID does not exist" containerID="ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.611020 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff"} err="failed to get container status \"ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff\": rpc error: code = NotFound desc = could not find container \"ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff\": container with ID starting with ecdf53f79e39c23ee6386f28b6b2ef19e116dd16782e6fedeac9b2ddbe88e2ff not found: ID does not exist" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.611037 4675 scope.go:117] "RemoveContainer" containerID="9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b" Feb 19 08:59:12 crc kubenswrapper[4675]: E0219 08:59:12.611275 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b\": container with ID starting with 9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b not found: ID does not exist" containerID="9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b" Feb 19 08:59:12 crc kubenswrapper[4675]: I0219 08:59:12.611301 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b"} err="failed to get container status \"9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b\": rpc error: code = NotFound desc = could not find container \"9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b\": container with ID starting with 9dbe5e45d3063b7d35009c47c46a70d47f8d2006349632cd10319e547bcf854b not found: ID does not exist" Feb 19 08:59:13 crc kubenswrapper[4675]: I0219 08:59:13.112003 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" path="/var/lib/kubelet/pods/b0ae0508-cd23-4122-bbe1-44686f753318/volumes" Feb 19 08:59:13 crc kubenswrapper[4675]: I0219 08:59:13.538105 4675 generic.go:334] "Generic (PLEG): container finished" podID="c5f740aa-7df7-4bb0-a388-1570876749df" containerID="d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600" exitCode=0 Feb 19 08:59:13 crc kubenswrapper[4675]: I0219 08:59:13.538207 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sv" event={"ID":"c5f740aa-7df7-4bb0-a388-1570876749df","Type":"ContainerDied","Data":"d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600"} Feb 19 08:59:14 crc kubenswrapper[4675]: I0219 08:59:14.553761 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sv" event={"ID":"c5f740aa-7df7-4bb0-a388-1570876749df","Type":"ContainerStarted","Data":"f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c"} Feb 19 08:59:14 crc kubenswrapper[4675]: I0219 08:59:14.573135 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6z8sv" podStartSLOduration=2.1295826 podStartE2EDuration="3.573118278s" podCreationTimestamp="2026-02-19 08:59:11 +0000 UTC" firstStartedPulling="2026-02-19 08:59:12.529782097 +0000 UTC m=+934.156872365" lastFinishedPulling="2026-02-19 08:59:13.973317775 +0000 UTC m=+935.600408043" observedRunningTime="2026-02-19 08:59:14.571063312 +0000 UTC m=+936.198153580" watchObservedRunningTime="2026-02-19 08:59:14.573118278 +0000 UTC m=+936.200208546" Feb 19 08:59:21 crc kubenswrapper[4675]: I0219 08:59:21.581323 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:21 crc kubenswrapper[4675]: I0219 08:59:21.581930 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:21 crc kubenswrapper[4675]: I0219 08:59:21.629486 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:21 crc kubenswrapper[4675]: I0219 08:59:21.671138 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:21 crc kubenswrapper[4675]: I0219 08:59:21.858615 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sv"] Feb 19 08:59:22 crc kubenswrapper[4675]: I0219 08:59:22.000006 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8" Feb 19 08:59:23 crc kubenswrapper[4675]: I0219 08:59:23.618424 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6z8sv" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="registry-server" containerID="cri-o://f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c" gracePeriod=2 Feb 19 08:59:23 crc kubenswrapper[4675]: I0219 08:59:23.995765 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.061082 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-catalog-content\") pod \"c5f740aa-7df7-4bb0-a388-1570876749df\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.061162 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-utilities\") pod \"c5f740aa-7df7-4bb0-a388-1570876749df\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.061234 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2bsl\" (UniqueName: \"kubernetes.io/projected/c5f740aa-7df7-4bb0-a388-1570876749df-kube-api-access-r2bsl\") pod \"c5f740aa-7df7-4bb0-a388-1570876749df\" (UID: \"c5f740aa-7df7-4bb0-a388-1570876749df\") " Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.062937 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-utilities" (OuterVolumeSpecName: "utilities") pod "c5f740aa-7df7-4bb0-a388-1570876749df" (UID: "c5f740aa-7df7-4bb0-a388-1570876749df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.068833 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5f740aa-7df7-4bb0-a388-1570876749df-kube-api-access-r2bsl" (OuterVolumeSpecName: "kube-api-access-r2bsl") pod "c5f740aa-7df7-4bb0-a388-1570876749df" (UID: "c5f740aa-7df7-4bb0-a388-1570876749df"). InnerVolumeSpecName "kube-api-access-r2bsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.089127 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5f740aa-7df7-4bb0-a388-1570876749df" (UID: "c5f740aa-7df7-4bb0-a388-1570876749df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.162832 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2bsl\" (UniqueName: \"kubernetes.io/projected/c5f740aa-7df7-4bb0-a388-1570876749df-kube-api-access-r2bsl\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.162866 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.162885 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5f740aa-7df7-4bb0-a388-1570876749df-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.467948 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fq49n"] Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.468492 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="extract-content" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468510 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="extract-content" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.468526 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="registry-server" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468533 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="registry-server" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.468552 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="extract-utilities" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468560 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="extract-utilities" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.468569 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="extract-content" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468576 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="extract-content" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.468590 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="extract-utilities" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468597 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="extract-utilities" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.468608 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="registry-server" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468614 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="registry-server" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468800 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ae0508-cd23-4122-bbe1-44686f753318" containerName="registry-server" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.468823 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" containerName="registry-server" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.472294 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.485651 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fq49n"] Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.570248 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-catalog-content\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.570369 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q86dv\" (UniqueName: \"kubernetes.io/projected/20744d3e-6767-48de-bb20-f12e9d2bc0cb-kube-api-access-q86dv\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.570396 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-utilities\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.625342 4675 generic.go:334] "Generic (PLEG): container finished" podID="c5f740aa-7df7-4bb0-a388-1570876749df" containerID="f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c" exitCode=0 Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.625405 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sv" event={"ID":"c5f740aa-7df7-4bb0-a388-1570876749df","Type":"ContainerDied","Data":"f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c"} Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.625455 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sv" event={"ID":"c5f740aa-7df7-4bb0-a388-1570876749df","Type":"ContainerDied","Data":"16da5ec17fb32efeed950f4489fc46a30587b5ea74723f3828cdbad4d90680a0"} Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.625474 4675 scope.go:117] "RemoveContainer" containerID="f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.627574 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sv" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.644291 4675 scope.go:117] "RemoveContainer" containerID="d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.659523 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sv"] Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.664190 4675 scope.go:117] "RemoveContainer" containerID="8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.666045 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sv"] Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.671176 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-catalog-content\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.671284 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q86dv\" (UniqueName: \"kubernetes.io/projected/20744d3e-6767-48de-bb20-f12e9d2bc0cb-kube-api-access-q86dv\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.671318 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-utilities\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.671820 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-utilities\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.672148 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-catalog-content\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.691422 4675 scope.go:117] "RemoveContainer" containerID="f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.692436 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q86dv\" (UniqueName: \"kubernetes.io/projected/20744d3e-6767-48de-bb20-f12e9d2bc0cb-kube-api-access-q86dv\") pod \"redhat-operators-fq49n\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.700596 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c\": container with ID starting with f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c not found: ID does not exist" containerID="f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.700678 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c"} err="failed to get container status \"f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c\": rpc error: code = NotFound desc = could not find container \"f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c\": container with ID starting with f23d84579a61f310e9f12d6a5405b09f4fd5c0e35e93a15db7517fe2eaba936c not found: ID does not exist" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.700707 4675 scope.go:117] "RemoveContainer" containerID="d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.702989 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600\": container with ID starting with d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600 not found: ID does not exist" containerID="d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.703045 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600"} err="failed to get container status \"d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600\": rpc error: code = NotFound desc = could not find container \"d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600\": container with ID starting with d1abbeec413ebd7306c34a285316e940431cceb04db771949faf047b7ad33600 not found: ID does not exist" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.703084 4675 scope.go:117] "RemoveContainer" containerID="8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a" Feb 19 08:59:24 crc kubenswrapper[4675]: E0219 08:59:24.703431 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a\": container with ID starting with 8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a not found: ID does not exist" containerID="8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.703463 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a"} err="failed to get container status \"8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a\": rpc error: code = NotFound desc = could not find container \"8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a\": container with ID starting with 8e5c090195467629dd14858650c02ef7caaf5ddf38087b5ed1424965d8dc0d3a not found: ID does not exist" Feb 19 08:59:24 crc kubenswrapper[4675]: I0219 08:59:24.797560 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:25 crc kubenswrapper[4675]: I0219 08:59:25.030396 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fq49n"] Feb 19 08:59:25 crc kubenswrapper[4675]: W0219 08:59:25.034449 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20744d3e_6767_48de_bb20_f12e9d2bc0cb.slice/crio-9cca55b053c733434556b1ee72d25ac398d688bd6afdf2c869eadcb189e38088 WatchSource:0}: Error finding container 9cca55b053c733434556b1ee72d25ac398d688bd6afdf2c869eadcb189e38088: Status 404 returned error can't find the container with id 9cca55b053c733434556b1ee72d25ac398d688bd6afdf2c869eadcb189e38088 Feb 19 08:59:25 crc kubenswrapper[4675]: I0219 08:59:25.115445 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5f740aa-7df7-4bb0-a388-1570876749df" path="/var/lib/kubelet/pods/c5f740aa-7df7-4bb0-a388-1570876749df/volumes" Feb 19 08:59:25 crc kubenswrapper[4675]: I0219 08:59:25.634107 4675 generic.go:334] "Generic (PLEG): container finished" podID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerID="6a4e58ef20e0cc30d0e71a8ea21da0e0ac60daf833c979076a698a207716a9b1" exitCode=0 Feb 19 08:59:25 crc kubenswrapper[4675]: I0219 08:59:25.634473 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq49n" event={"ID":"20744d3e-6767-48de-bb20-f12e9d2bc0cb","Type":"ContainerDied","Data":"6a4e58ef20e0cc30d0e71a8ea21da0e0ac60daf833c979076a698a207716a9b1"} Feb 19 08:59:25 crc kubenswrapper[4675]: I0219 08:59:25.634500 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq49n" event={"ID":"20744d3e-6767-48de-bb20-f12e9d2bc0cb","Type":"ContainerStarted","Data":"9cca55b053c733434556b1ee72d25ac398d688bd6afdf2c869eadcb189e38088"} Feb 19 08:59:27 crc kubenswrapper[4675]: I0219 08:59:27.652849 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq49n" event={"ID":"20744d3e-6767-48de-bb20-f12e9d2bc0cb","Type":"ContainerStarted","Data":"15592f8d53cdc72f7ca26c4364678119b75a3c811062dfd98c0e30e1cac705d4"} Feb 19 08:59:28 crc kubenswrapper[4675]: I0219 08:59:28.662129 4675 generic.go:334] "Generic (PLEG): container finished" podID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerID="15592f8d53cdc72f7ca26c4364678119b75a3c811062dfd98c0e30e1cac705d4" exitCode=0 Feb 19 08:59:28 crc kubenswrapper[4675]: I0219 08:59:28.662178 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq49n" event={"ID":"20744d3e-6767-48de-bb20-f12e9d2bc0cb","Type":"ContainerDied","Data":"15592f8d53cdc72f7ca26c4364678119b75a3c811062dfd98c0e30e1cac705d4"} Feb 19 08:59:29 crc kubenswrapper[4675]: I0219 08:59:29.672508 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq49n" event={"ID":"20744d3e-6767-48de-bb20-f12e9d2bc0cb","Type":"ContainerStarted","Data":"49f34493c3a8805fca63cfe9fa2f217a0d9cd079b800272e754008f8ded0aa9d"} Feb 19 08:59:29 crc kubenswrapper[4675]: I0219 08:59:29.692358 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fq49n" podStartSLOduration=2.2608959779999998 podStartE2EDuration="5.69233566s" podCreationTimestamp="2026-02-19 08:59:24 +0000 UTC" firstStartedPulling="2026-02-19 08:59:25.635777992 +0000 UTC m=+947.262868260" lastFinishedPulling="2026-02-19 08:59:29.067217674 +0000 UTC m=+950.694307942" observedRunningTime="2026-02-19 08:59:29.69050624 +0000 UTC m=+951.317596538" watchObservedRunningTime="2026-02-19 08:59:29.69233566 +0000 UTC m=+951.319425928" Feb 19 08:59:34 crc kubenswrapper[4675]: I0219 08:59:34.797671 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:34 crc kubenswrapper[4675]: I0219 08:59:34.798189 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:34 crc kubenswrapper[4675]: I0219 08:59:34.839622 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:35 crc kubenswrapper[4675]: I0219 08:59:35.766612 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:35 crc kubenswrapper[4675]: I0219 08:59:35.812032 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fq49n"] Feb 19 08:59:37 crc kubenswrapper[4675]: I0219 08:59:37.736113 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fq49n" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="registry-server" containerID="cri-o://49f34493c3a8805fca63cfe9fa2f217a0d9cd079b800272e754008f8ded0aa9d" gracePeriod=2 Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.548799 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vzbdl"] Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.550287 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.553381 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.553631 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5rxn6" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.555177 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.555409 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.561390 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vzbdl"] Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.626796 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d9gkd"] Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.628097 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.630347 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.634523 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d9gkd"] Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.659199 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m45tw\" (UniqueName: \"kubernetes.io/projected/156bcac8-d2d5-4ba0-819a-aa0f253e378c-kube-api-access-m45tw\") pod \"dnsmasq-dns-675f4bcbfc-vzbdl\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.659236 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.659387 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bhbw\" (UniqueName: \"kubernetes.io/projected/489b62db-1897-4b8b-809a-6d2906775d93-kube-api-access-9bhbw\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.659440 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bcac8-d2d5-4ba0-819a-aa0f253e378c-config\") pod \"dnsmasq-dns-675f4bcbfc-vzbdl\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.659492 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-config\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.746511 4675 generic.go:334] "Generic (PLEG): container finished" podID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerID="49f34493c3a8805fca63cfe9fa2f217a0d9cd079b800272e754008f8ded0aa9d" exitCode=0 Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.746553 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq49n" event={"ID":"20744d3e-6767-48de-bb20-f12e9d2bc0cb","Type":"ContainerDied","Data":"49f34493c3a8805fca63cfe9fa2f217a0d9cd079b800272e754008f8ded0aa9d"} Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.761300 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bhbw\" (UniqueName: \"kubernetes.io/projected/489b62db-1897-4b8b-809a-6d2906775d93-kube-api-access-9bhbw\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.761358 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bcac8-d2d5-4ba0-819a-aa0f253e378c-config\") pod \"dnsmasq-dns-675f4bcbfc-vzbdl\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.761389 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-config\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.761426 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m45tw\" (UniqueName: \"kubernetes.io/projected/156bcac8-d2d5-4ba0-819a-aa0f253e378c-kube-api-access-m45tw\") pod \"dnsmasq-dns-675f4bcbfc-vzbdl\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.761441 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.762357 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.762370 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bcac8-d2d5-4ba0-819a-aa0f253e378c-config\") pod \"dnsmasq-dns-675f4bcbfc-vzbdl\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.762390 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-config\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.783514 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bhbw\" (UniqueName: \"kubernetes.io/projected/489b62db-1897-4b8b-809a-6d2906775d93-kube-api-access-9bhbw\") pod \"dnsmasq-dns-78dd6ddcc-d9gkd\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.792310 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m45tw\" (UniqueName: \"kubernetes.io/projected/156bcac8-d2d5-4ba0-819a-aa0f253e378c-kube-api-access-m45tw\") pod \"dnsmasq-dns-675f4bcbfc-vzbdl\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.928273 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 08:59:38 crc kubenswrapper[4675]: I0219 08:59:38.944013 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.366422 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.469278 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-catalog-content\") pod \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.469773 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q86dv\" (UniqueName: \"kubernetes.io/projected/20744d3e-6767-48de-bb20-f12e9d2bc0cb-kube-api-access-q86dv\") pod \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.469837 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-utilities\") pod \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\" (UID: \"20744d3e-6767-48de-bb20-f12e9d2bc0cb\") " Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.471421 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-utilities" (OuterVolumeSpecName: "utilities") pod "20744d3e-6767-48de-bb20-f12e9d2bc0cb" (UID: "20744d3e-6767-48de-bb20-f12e9d2bc0cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.477963 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20744d3e-6767-48de-bb20-f12e9d2bc0cb-kube-api-access-q86dv" (OuterVolumeSpecName: "kube-api-access-q86dv") pod "20744d3e-6767-48de-bb20-f12e9d2bc0cb" (UID: "20744d3e-6767-48de-bb20-f12e9d2bc0cb"). InnerVolumeSpecName "kube-api-access-q86dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.517108 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d9gkd"] Feb 19 08:59:39 crc kubenswrapper[4675]: W0219 08:59:39.520553 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod489b62db_1897_4b8b_809a_6d2906775d93.slice/crio-6d248e18472cfeb7d2bb7852d3e1317ddcd3f4e1c7880a984679851bf7113bed WatchSource:0}: Error finding container 6d248e18472cfeb7d2bb7852d3e1317ddcd3f4e1c7880a984679851bf7113bed: Status 404 returned error can't find the container with id 6d248e18472cfeb7d2bb7852d3e1317ddcd3f4e1c7880a984679851bf7113bed Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.522263 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vzbdl"] Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.523783 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.574292 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q86dv\" (UniqueName: \"kubernetes.io/projected/20744d3e-6767-48de-bb20-f12e9d2bc0cb-kube-api-access-q86dv\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.574336 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.624654 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20744d3e-6767-48de-bb20-f12e9d2bc0cb" (UID: "20744d3e-6767-48de-bb20-f12e9d2bc0cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.676010 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20744d3e-6767-48de-bb20-f12e9d2bc0cb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.754451 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fq49n" event={"ID":"20744d3e-6767-48de-bb20-f12e9d2bc0cb","Type":"ContainerDied","Data":"9cca55b053c733434556b1ee72d25ac398d688bd6afdf2c869eadcb189e38088"} Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.754472 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fq49n" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.754517 4675 scope.go:117] "RemoveContainer" containerID="49f34493c3a8805fca63cfe9fa2f217a0d9cd079b800272e754008f8ded0aa9d" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.755682 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" event={"ID":"156bcac8-d2d5-4ba0-819a-aa0f253e378c","Type":"ContainerStarted","Data":"f066db28f76aaef45938da64932c5c32c17e1eea421f54fe3d09d90aa78ddd11"} Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.757592 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" event={"ID":"489b62db-1897-4b8b-809a-6d2906775d93","Type":"ContainerStarted","Data":"6d248e18472cfeb7d2bb7852d3e1317ddcd3f4e1c7880a984679851bf7113bed"} Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.775837 4675 scope.go:117] "RemoveContainer" containerID="15592f8d53cdc72f7ca26c4364678119b75a3c811062dfd98c0e30e1cac705d4" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.812153 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fq49n"] Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.813139 4675 scope.go:117] "RemoveContainer" containerID="6a4e58ef20e0cc30d0e71a8ea21da0e0ac60daf833c979076a698a207716a9b1" Feb 19 08:59:39 crc kubenswrapper[4675]: I0219 08:59:39.820846 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fq49n"] Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.126067 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" path="/var/lib/kubelet/pods/20744d3e-6767-48de-bb20-f12e9d2bc0cb/volumes" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.258171 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vzbdl"] Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.287263 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nnzpk"] Feb 19 08:59:41 crc kubenswrapper[4675]: E0219 08:59:41.287602 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="extract-utilities" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.287622 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="extract-utilities" Feb 19 08:59:41 crc kubenswrapper[4675]: E0219 08:59:41.287654 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="extract-content" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.287663 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="extract-content" Feb 19 08:59:41 crc kubenswrapper[4675]: E0219 08:59:41.287681 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="registry-server" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.287688 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="registry-server" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.287840 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="20744d3e-6767-48de-bb20-f12e9d2bc0cb" containerName="registry-server" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.288616 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.296403 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nnzpk"] Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.417586 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-config\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.417677 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmf9n\" (UniqueName: \"kubernetes.io/projected/dc6d527a-29ec-432a-a7c1-87a54af36db8-kube-api-access-wmf9n\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.417722 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.519454 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.519543 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-config\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.519589 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmf9n\" (UniqueName: \"kubernetes.io/projected/dc6d527a-29ec-432a-a7c1-87a54af36db8-kube-api-access-wmf9n\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.520443 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.521258 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-config\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.554994 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmf9n\" (UniqueName: \"kubernetes.io/projected/dc6d527a-29ec-432a-a7c1-87a54af36db8-kube-api-access-wmf9n\") pod \"dnsmasq-dns-666b6646f7-nnzpk\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.561125 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d9gkd"] Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.589784 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chnjb"] Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.591178 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.604792 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chnjb"] Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.614600 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.723720 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-config\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.723822 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.723873 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7jb7\" (UniqueName: \"kubernetes.io/projected/a5a4aa24-a744-45fb-88a6-546affb72d83-kube-api-access-z7jb7\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.828162 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.828685 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7jb7\" (UniqueName: \"kubernetes.io/projected/a5a4aa24-a744-45fb-88a6-546affb72d83-kube-api-access-z7jb7\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.828779 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-config\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.829474 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.829688 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-config\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.877586 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7jb7\" (UniqueName: \"kubernetes.io/projected/a5a4aa24-a744-45fb-88a6-546affb72d83-kube-api-access-z7jb7\") pod \"dnsmasq-dns-57d769cc4f-chnjb\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:41 crc kubenswrapper[4675]: I0219 08:59:41.909680 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.193765 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nnzpk"] Feb 19 08:59:42 crc kubenswrapper[4675]: W0219 08:59:42.206759 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc6d527a_29ec_432a_a7c1_87a54af36db8.slice/crio-f8334309e751429de474f4a441cf69cdcafcd463ab19354d509a0c4e3868b5e6 WatchSource:0}: Error finding container f8334309e751429de474f4a441cf69cdcafcd463ab19354d509a0c4e3868b5e6: Status 404 returned error can't find the container with id f8334309e751429de474f4a441cf69cdcafcd463ab19354d509a0c4e3868b5e6 Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.441162 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.442564 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.449703 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.449808 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.449958 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.450026 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.450098 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.450211 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.450252 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-f4b2m" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.473655 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chnjb"] Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.503169 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.543650 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93dbb123-819d-44f7-80f5-8c38c639f5f4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.543724 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.543756 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.543804 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.543832 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93dbb123-819d-44f7-80f5-8c38c639f5f4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.543876 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.543899 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.544015 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.544091 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.544156 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bxqz\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-kube-api-access-6bxqz\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.544208 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-config-data\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651366 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651417 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651440 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651464 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651495 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bxqz\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-kube-api-access-6bxqz\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651519 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-config-data\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651539 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93dbb123-819d-44f7-80f5-8c38c639f5f4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651560 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651580 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651612 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.651630 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93dbb123-819d-44f7-80f5-8c38c639f5f4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.653038 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.653254 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-config-data\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.654120 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.655016 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.655113 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.655951 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.670375 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93dbb123-819d-44f7-80f5-8c38c639f5f4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.670551 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.678958 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93dbb123-819d-44f7-80f5-8c38c639f5f4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.681434 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.688270 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bxqz\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-kube-api-access-6bxqz\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.701055 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.723908 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.742631 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.746872 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.746980 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.747196 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-88zj9" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.747245 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.747372 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.747486 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.749146 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.766395 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.795999 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.804719 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" event={"ID":"dc6d527a-29ec-432a-a7c1-87a54af36db8","Type":"ContainerStarted","Data":"f8334309e751429de474f4a441cf69cdcafcd463ab19354d509a0c4e3868b5e6"} Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.806011 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" event={"ID":"a5a4aa24-a744-45fb-88a6-546affb72d83","Type":"ContainerStarted","Data":"3d5074c9c648dac25d6f24ca87868c6df8b10a78d999d5418062c87e5cc73041"} Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.859921 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860207 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860231 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1adde38f-79d6-4ee4-82ed-97004ffb7f85-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860398 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860424 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860453 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rnc5\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-kube-api-access-9rnc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860483 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860510 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860527 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1adde38f-79d6-4ee4-82ed-97004ffb7f85-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860565 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.860585 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983463 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983508 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1adde38f-79d6-4ee4-82ed-97004ffb7f85-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983540 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983563 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983588 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rnc5\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-kube-api-access-9rnc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983617 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983654 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983671 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1adde38f-79d6-4ee4-82ed-97004ffb7f85-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983706 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983723 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.983743 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.985282 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.985520 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:42 crc kubenswrapper[4675]: I0219 08:59:42.991172 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:42.998751 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:42.999400 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1adde38f-79d6-4ee4-82ed-97004ffb7f85-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:42.999451 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:42.999507 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:42.999634 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1adde38f-79d6-4ee4-82ed-97004ffb7f85-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.003733 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.004272 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.015709 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rnc5\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-kube-api-access-9rnc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.018878 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.104632 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.346805 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 08:59:43 crc kubenswrapper[4675]: W0219 08:59:43.394224 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93dbb123_819d_44f7_80f5_8c38c639f5f4.slice/crio-b38e71e5de19bb8f73b6cbd485738542f19363ab0d2195391d0b2a32a3a04cb8 WatchSource:0}: Error finding container b38e71e5de19bb8f73b6cbd485738542f19363ab0d2195391d0b2a32a3a04cb8: Status 404 returned error can't find the container with id b38e71e5de19bb8f73b6cbd485738542f19363ab0d2195391d0b2a32a3a04cb8 Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.616302 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 08:59:43 crc kubenswrapper[4675]: W0219 08:59:43.649584 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1adde38f_79d6_4ee4_82ed_97004ffb7f85.slice/crio-5c1fd77cea3aa2e47cce7e46f2baf1ed9a93d610455260f098319db176372cf4 WatchSource:0}: Error finding container 5c1fd77cea3aa2e47cce7e46f2baf1ed9a93d610455260f098319db176372cf4: Status 404 returned error can't find the container with id 5c1fd77cea3aa2e47cce7e46f2baf1ed9a93d610455260f098319db176372cf4 Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.768717 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.770094 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.772444 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.780728 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4ljmz" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.778128 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.783078 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.789134 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.801202 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.846743 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1adde38f-79d6-4ee4-82ed-97004ffb7f85","Type":"ContainerStarted","Data":"5c1fd77cea3aa2e47cce7e46f2baf1ed9a93d610455260f098319db176372cf4"} Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.851457 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93dbb123-819d-44f7-80f5-8c38c639f5f4","Type":"ContainerStarted","Data":"b38e71e5de19bb8f73b6cbd485738542f19363ab0d2195391d0b2a32a3a04cb8"} Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.904742 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.904842 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh9bl\" (UniqueName: \"kubernetes.io/projected/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-kube-api-access-fh9bl\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.904890 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.904964 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.905177 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-config-data-default\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.905290 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.905322 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:43 crc kubenswrapper[4675]: I0219 08:59:43.905533 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-kolla-config\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009083 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009160 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-config-data-default\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009187 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009209 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009272 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-kolla-config\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009313 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009342 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh9bl\" (UniqueName: \"kubernetes.io/projected/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-kube-api-access-fh9bl\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009376 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.009931 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.010751 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.012583 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-kolla-config\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.013573 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-config-data-default\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.013814 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.020892 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.021450 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.043703 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh9bl\" (UniqueName: \"kubernetes.io/projected/b91a4d30-10a1-4827-b4a7-2c620a2c2d9f-kube-api-access-fh9bl\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.059295 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f\") " pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.118860 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.749248 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 19 08:59:44 crc kubenswrapper[4675]: I0219 08:59:44.866718 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f","Type":"ContainerStarted","Data":"836073f4e8dbd00aabd164efacdf8a9aa7b01b29c4c0612230eb88bfda92d6f4"} Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.100726 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.205253 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.211734 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.217682 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.224140 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-lhrkq" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.224855 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.225165 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233657 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b301f22-5f49-490c-899a-f518d8da2174-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233709 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233732 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b301f22-5f49-490c-899a-f518d8da2174-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233754 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233844 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b301f22-5f49-490c-899a-f518d8da2174-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233880 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l7d7\" (UniqueName: \"kubernetes.io/projected/3b301f22-5f49-490c-899a-f518d8da2174-kube-api-access-2l7d7\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233897 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.233932 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.334815 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b301f22-5f49-490c-899a-f518d8da2174-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.334875 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.334898 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b301f22-5f49-490c-899a-f518d8da2174-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.334916 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.334945 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b301f22-5f49-490c-899a-f518d8da2174-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.334975 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l7d7\" (UniqueName: \"kubernetes.io/projected/3b301f22-5f49-490c-899a-f518d8da2174-kube-api-access-2l7d7\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.334992 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.335867 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.336049 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.336495 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b301f22-5f49-490c-899a-f518d8da2174-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.336747 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.337778 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.344606 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b301f22-5f49-490c-899a-f518d8da2174-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.352877 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b301f22-5f49-490c-899a-f518d8da2174-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.362955 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l7d7\" (UniqueName: \"kubernetes.io/projected/3b301f22-5f49-490c-899a-f518d8da2174-kube-api-access-2l7d7\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.372411 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.374957 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b301f22-5f49-490c-899a-f518d8da2174-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3b301f22-5f49-490c-899a-f518d8da2174\") " pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.496155 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.497387 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.501138 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.503035 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.509975 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.510471 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-fzp84" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.550688 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.640376 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcr9c\" (UniqueName: \"kubernetes.io/projected/4f51af02-6352-4462-b00f-a2feafe19f23-kube-api-access-pcr9c\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.640527 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f51af02-6352-4462-b00f-a2feafe19f23-config-data\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.640561 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f51af02-6352-4462-b00f-a2feafe19f23-kolla-config\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.640620 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f51af02-6352-4462-b00f-a2feafe19f23-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.641529 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f51af02-6352-4462-b00f-a2feafe19f23-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.742986 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcr9c\" (UniqueName: \"kubernetes.io/projected/4f51af02-6352-4462-b00f-a2feafe19f23-kube-api-access-pcr9c\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.743035 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f51af02-6352-4462-b00f-a2feafe19f23-config-data\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.743076 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f51af02-6352-4462-b00f-a2feafe19f23-kolla-config\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.743113 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f51af02-6352-4462-b00f-a2feafe19f23-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.743156 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f51af02-6352-4462-b00f-a2feafe19f23-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.744264 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f51af02-6352-4462-b00f-a2feafe19f23-config-data\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.744278 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f51af02-6352-4462-b00f-a2feafe19f23-kolla-config\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.767544 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f51af02-6352-4462-b00f-a2feafe19f23-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.769620 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcr9c\" (UniqueName: \"kubernetes.io/projected/4f51af02-6352-4462-b00f-a2feafe19f23-kube-api-access-pcr9c\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:45 crc kubenswrapper[4675]: I0219 08:59:45.770052 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f51af02-6352-4462-b00f-a2feafe19f23-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4f51af02-6352-4462-b00f-a2feafe19f23\") " pod="openstack/memcached-0" Feb 19 08:59:46 crc kubenswrapper[4675]: I0219 08:59:45.832756 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.620959 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.622377 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.635464 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.644917 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-stb5l" Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.784149 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnhw4\" (UniqueName: \"kubernetes.io/projected/ba2e218a-edb2-4626-b480-1e4a534bc80d-kube-api-access-cnhw4\") pod \"kube-state-metrics-0\" (UID: \"ba2e218a-edb2-4626-b480-1e4a534bc80d\") " pod="openstack/kube-state-metrics-0" Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.885806 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnhw4\" (UniqueName: \"kubernetes.io/projected/ba2e218a-edb2-4626-b480-1e4a534bc80d-kube-api-access-cnhw4\") pod \"kube-state-metrics-0\" (UID: \"ba2e218a-edb2-4626-b480-1e4a534bc80d\") " pod="openstack/kube-state-metrics-0" Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.912106 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnhw4\" (UniqueName: \"kubernetes.io/projected/ba2e218a-edb2-4626-b480-1e4a534bc80d-kube-api-access-cnhw4\") pod \"kube-state-metrics-0\" (UID: \"ba2e218a-edb2-4626-b480-1e4a534bc80d\") " pod="openstack/kube-state-metrics-0" Feb 19 08:59:47 crc kubenswrapper[4675]: I0219 08:59:47.953710 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.681409 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vlnhc"] Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.683089 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.685685 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-25g87" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.685816 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.686208 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.705869 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vlnhc"] Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.740524 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b11fc83-5e56-43d3-a340-80686a5bdf35-scripts\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.740578 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b11fc83-5e56-43d3-a340-80686a5bdf35-ovn-controller-tls-certs\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.740599 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xh9g\" (UniqueName: \"kubernetes.io/projected/0b11fc83-5e56-43d3-a340-80686a5bdf35-kube-api-access-8xh9g\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.740650 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-run\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.740684 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b11fc83-5e56-43d3-a340-80686a5bdf35-combined-ca-bundle\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.740725 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-run-ovn\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.740745 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-log-ovn\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.800252 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-s5dht"] Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.801826 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.820588 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-s5dht"] Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843525 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xh9g\" (UniqueName: \"kubernetes.io/projected/0b11fc83-5e56-43d3-a340-80686a5bdf35-kube-api-access-8xh9g\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843584 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-etc-ovs\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843615 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-run\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843662 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b11fc83-5e56-43d3-a340-80686a5bdf35-combined-ca-bundle\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843692 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zqvq\" (UniqueName: \"kubernetes.io/projected/6dd9438b-4fad-4a66-8d08-df4816a19124-kube-api-access-2zqvq\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843712 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-log\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843735 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-run\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843755 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-run-ovn\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843778 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-log-ovn\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843814 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-lib\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843837 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b11fc83-5e56-43d3-a340-80686a5bdf35-scripts\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843853 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6dd9438b-4fad-4a66-8d08-df4816a19124-scripts\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.843870 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b11fc83-5e56-43d3-a340-80686a5bdf35-ovn-controller-tls-certs\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.845037 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-run\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.845296 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-log-ovn\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.845448 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0b11fc83-5e56-43d3-a340-80686a5bdf35-var-run-ovn\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.847186 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b11fc83-5e56-43d3-a340-80686a5bdf35-scripts\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.856243 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b11fc83-5e56-43d3-a340-80686a5bdf35-ovn-controller-tls-certs\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.872343 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b11fc83-5e56-43d3-a340-80686a5bdf35-combined-ca-bundle\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.889437 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xh9g\" (UniqueName: \"kubernetes.io/projected/0b11fc83-5e56-43d3-a340-80686a5bdf35-kube-api-access-8xh9g\") pod \"ovn-controller-vlnhc\" (UID: \"0b11fc83-5e56-43d3-a340-80686a5bdf35\") " pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.945382 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zqvq\" (UniqueName: \"kubernetes.io/projected/6dd9438b-4fad-4a66-8d08-df4816a19124-kube-api-access-2zqvq\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.945729 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-log\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.945873 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-run\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.945995 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-lib\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.946097 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6dd9438b-4fad-4a66-8d08-df4816a19124-scripts\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.946176 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-etc-ovs\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.946228 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-run\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.946239 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-log\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.946360 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-var-lib\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.946948 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6dd9438b-4fad-4a66-8d08-df4816a19124-etc-ovs\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.955237 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6dd9438b-4fad-4a66-8d08-df4816a19124-scripts\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:50 crc kubenswrapper[4675]: I0219 08:59:50.983844 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zqvq\" (UniqueName: \"kubernetes.io/projected/6dd9438b-4fad-4a66-8d08-df4816a19124-kube-api-access-2zqvq\") pod \"ovn-controller-ovs-s5dht\" (UID: \"6dd9438b-4fad-4a66-8d08-df4816a19124\") " pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:51 crc kubenswrapper[4675]: I0219 08:59:51.009227 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc" Feb 19 08:59:51 crc kubenswrapper[4675]: I0219 08:59:51.283849 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-s5dht" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.087170 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.088763 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.093622 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-ntx6w" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.093815 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.093844 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.096027 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.096487 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.122927 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.287915 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.287974 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.289038 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.289150 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2344b53c-5df6-4673-beff-7c3cdd28dc93-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.289266 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7f64\" (UniqueName: \"kubernetes.io/projected/2344b53c-5df6-4673-beff-7c3cdd28dc93-kube-api-access-d7f64\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.289297 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2344b53c-5df6-4673-beff-7c3cdd28dc93-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.289326 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2344b53c-5df6-4673-beff-7c3cdd28dc93-config\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.289707 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391173 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391235 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391295 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391352 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2344b53c-5df6-4673-beff-7c3cdd28dc93-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391412 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7f64\" (UniqueName: \"kubernetes.io/projected/2344b53c-5df6-4673-beff-7c3cdd28dc93-kube-api-access-d7f64\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391435 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2344b53c-5df6-4673-beff-7c3cdd28dc93-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391465 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2344b53c-5df6-4673-beff-7c3cdd28dc93-config\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.391486 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.393135 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2344b53c-5df6-4673-beff-7c3cdd28dc93-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.393783 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2344b53c-5df6-4673-beff-7c3cdd28dc93-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.393915 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.395965 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2344b53c-5df6-4673-beff-7c3cdd28dc93-config\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.397826 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.398608 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.408634 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2344b53c-5df6-4673-beff-7c3cdd28dc93-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.411686 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7f64\" (UniqueName: \"kubernetes.io/projected/2344b53c-5df6-4673-beff-7c3cdd28dc93-kube-api-access-d7f64\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.418635 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2344b53c-5df6-4673-beff-7c3cdd28dc93\") " pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:53 crc kubenswrapper[4675]: I0219 08:59:53.427157 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 19 08:59:54 crc kubenswrapper[4675]: I0219 08:59:54.892959 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 19 08:59:54 crc kubenswrapper[4675]: I0219 08:59:54.894808 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:54 crc kubenswrapper[4675]: I0219 08:59:54.906871 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-f4kbs" Feb 19 08:59:54 crc kubenswrapper[4675]: I0219 08:59:54.907444 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 19 08:59:54 crc kubenswrapper[4675]: I0219 08:59:54.907668 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 19 08:59:54 crc kubenswrapper[4675]: I0219 08:59:54.907855 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 19 08:59:54 crc kubenswrapper[4675]: I0219 08:59:54.918124 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.027955 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.028002 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-config\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.028097 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttj7x\" (UniqueName: \"kubernetes.io/projected/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-kube-api-access-ttj7x\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.028117 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.028153 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.028178 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.028206 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.028955 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.131601 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.131688 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.131719 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.131754 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.132930 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.132938 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.132982 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-config\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.133096 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.133122 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttj7x\" (UniqueName: \"kubernetes.io/projected/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-kube-api-access-ttj7x\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.133912 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.134152 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.135181 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-config\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.146631 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.146896 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.147511 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.150087 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttj7x\" (UniqueName: \"kubernetes.io/projected/482efc7c-8e44-4c25-903e-5c8b2cc58d0b-kube-api-access-ttj7x\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.160276 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"482efc7c-8e44-4c25-903e-5c8b2cc58d0b\") " pod="openstack/ovsdbserver-nb-0" Feb 19 08:59:55 crc kubenswrapper[4675]: I0219 08:59:55.223789 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.150362 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8"] Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.151754 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.156006 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.156032 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.163969 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8"] Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.322240 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75kww\" (UniqueName: \"kubernetes.io/projected/7a68599d-2cc4-4323-8437-58d0866e994e-kube-api-access-75kww\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.322535 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a68599d-2cc4-4323-8437-58d0866e994e-secret-volume\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.323184 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a68599d-2cc4-4323-8437-58d0866e994e-config-volume\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.424214 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a68599d-2cc4-4323-8437-58d0866e994e-config-volume\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.425028 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a68599d-2cc4-4323-8437-58d0866e994e-config-volume\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.425113 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75kww\" (UniqueName: \"kubernetes.io/projected/7a68599d-2cc4-4323-8437-58d0866e994e-kube-api-access-75kww\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.425253 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a68599d-2cc4-4323-8437-58d0866e994e-secret-volume\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.440469 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a68599d-2cc4-4323-8437-58d0866e994e-secret-volume\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.447848 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75kww\" (UniqueName: \"kubernetes.io/projected/7a68599d-2cc4-4323-8437-58d0866e994e-kube-api-access-75kww\") pod \"collect-profiles-29524860-mf7w8\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:00 crc kubenswrapper[4675]: I0219 09:00:00.475469 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:03 crc kubenswrapper[4675]: E0219 09:00:03.099314 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Feb 19 09:00:03 crc kubenswrapper[4675]: E0219 09:00:03.099856 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fh9bl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(b91a4d30-10a1-4827-b4a7-2c620a2c2d9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:00:03 crc kubenswrapper[4675]: E0219 09:00:03.101053 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="b91a4d30-10a1-4827-b4a7-2c620a2c2d9f" Feb 19 09:00:03 crc kubenswrapper[4675]: E0219 09:00:03.121676 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Feb 19 09:00:03 crc kubenswrapper[4675]: E0219 09:00:03.121845 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6bxqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(93dbb123-819d-44f7-80f5-8c38c639f5f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:00:03 crc kubenswrapper[4675]: E0219 09:00:03.123621 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" Feb 19 09:00:04 crc kubenswrapper[4675]: E0219 09:00:04.031088 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="b91a4d30-10a1-4827-b4a7-2c620a2c2d9f" Feb 19 09:00:04 crc kubenswrapper[4675]: E0219 09:00:04.034283 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" Feb 19 09:00:10 crc kubenswrapper[4675]: I0219 09:00:10.720146 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 09:00:10 crc kubenswrapper[4675]: I0219 09:00:10.810631 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 19 09:00:10 crc kubenswrapper[4675]: I0219 09:00:10.902528 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-s5dht"] Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.319187 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.319767 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m45tw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-vzbdl_openstack(156bcac8-d2d5-4ba0-819a-aa0f253e378c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.321177 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" podUID="156bcac8-d2d5-4ba0-819a-aa0f253e378c" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.331651 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.331816 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wmf9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-nnzpk_openstack(dc6d527a-29ec-432a-a7c1-87a54af36db8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.335747 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" podUID="dc6d527a-29ec-432a-a7c1-87a54af36db8" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.373852 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.374045 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9bhbw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-d9gkd_openstack(489b62db-1897-4b8b-809a-6d2906775d93): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.374326 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.374401 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z7jb7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-chnjb_openstack(a5a4aa24-a744-45fb-88a6-546affb72d83): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.382596 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" podUID="a5a4aa24-a744-45fb-88a6-546affb72d83" Feb 19 09:00:11 crc kubenswrapper[4675]: E0219 09:00:11.382706 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" podUID="489b62db-1897-4b8b-809a-6d2906775d93" Feb 19 09:00:11 crc kubenswrapper[4675]: I0219 09:00:11.644657 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vlnhc"] Feb 19 09:00:11 crc kubenswrapper[4675]: I0219 09:00:11.743945 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:00:11 crc kubenswrapper[4675]: I0219 09:00:11.743998 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:00:11 crc kubenswrapper[4675]: I0219 09:00:11.912529 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8"] Feb 19 09:00:11 crc kubenswrapper[4675]: I0219 09:00:11.926513 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 19 09:00:11 crc kubenswrapper[4675]: I0219 09:00:11.990491 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.038496 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 19 09:00:12 crc kubenswrapper[4675]: W0219 09:00:12.061276 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b301f22_5f49_490c_899a_f518d8da2174.slice/crio-9e45cf294c988395bfa930c357e9f5aaa0738f2755af8123c17699c5aa4d85c0 WatchSource:0}: Error finding container 9e45cf294c988395bfa930c357e9f5aaa0738f2755af8123c17699c5aa4d85c0: Status 404 returned error can't find the container with id 9e45cf294c988395bfa930c357e9f5aaa0738f2755af8123c17699c5aa4d85c0 Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.112721 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" event={"ID":"7a68599d-2cc4-4323-8437-58d0866e994e","Type":"ContainerStarted","Data":"7583ac43950a14c75c4da27823130b38321389535d0104976dfedeee43c00607"} Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.114114 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3b301f22-5f49-490c-899a-f518d8da2174","Type":"ContainerStarted","Data":"9e45cf294c988395bfa930c357e9f5aaa0738f2755af8123c17699c5aa4d85c0"} Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.115272 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4f51af02-6352-4462-b00f-a2feafe19f23","Type":"ContainerStarted","Data":"8210f86043cc02fb6798c377fe2f7d6c930f276665048138adf20d7e044dbea1"} Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.116542 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ba2e218a-edb2-4626-b480-1e4a534bc80d","Type":"ContainerStarted","Data":"7b32ea5c053a87da0165bf77be36ad6a13bdbef9c109973347a38b206c0e99aa"} Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.117846 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2344b53c-5df6-4673-beff-7c3cdd28dc93","Type":"ContainerStarted","Data":"bafb2f9960fdaf452cd7f54b15bc151a6c34100141f11d62655ad930491825d9"} Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.118975 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"482efc7c-8e44-4c25-903e-5c8b2cc58d0b","Type":"ContainerStarted","Data":"e602c661b23a96e693d5c88a0832b6942341cda1086a2e83bbaa709cfb095994"} Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.120243 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5dht" event={"ID":"6dd9438b-4fad-4a66-8d08-df4816a19124","Type":"ContainerStarted","Data":"ffa31b2fae4b574bb83831e65c992f45db7fc2633a5bdab867ce9428266d22b9"} Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.121222 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc" event={"ID":"0b11fc83-5e56-43d3-a340-80686a5bdf35","Type":"ContainerStarted","Data":"394ab4c7bfad56412f3ffa62aeef418630c2d43baeffbd22325bbc845a6d0008"} Feb 19 09:00:12 crc kubenswrapper[4675]: E0219 09:00:12.122815 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" podUID="dc6d527a-29ec-432a-a7c1-87a54af36db8" Feb 19 09:00:12 crc kubenswrapper[4675]: E0219 09:00:12.122961 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" podUID="a5a4aa24-a744-45fb-88a6-546affb72d83" Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.797166 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.802594 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.974317 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bcac8-d2d5-4ba0-819a-aa0f253e378c-config\") pod \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.974390 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-config\") pod \"489b62db-1897-4b8b-809a-6d2906775d93\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.974421 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bhbw\" (UniqueName: \"kubernetes.io/projected/489b62db-1897-4b8b-809a-6d2906775d93-kube-api-access-9bhbw\") pod \"489b62db-1897-4b8b-809a-6d2906775d93\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.974518 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m45tw\" (UniqueName: \"kubernetes.io/projected/156bcac8-d2d5-4ba0-819a-aa0f253e378c-kube-api-access-m45tw\") pod \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\" (UID: \"156bcac8-d2d5-4ba0-819a-aa0f253e378c\") " Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.974535 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-dns-svc\") pod \"489b62db-1897-4b8b-809a-6d2906775d93\" (UID: \"489b62db-1897-4b8b-809a-6d2906775d93\") " Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.975123 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156bcac8-d2d5-4ba0-819a-aa0f253e378c-config" (OuterVolumeSpecName: "config") pod "156bcac8-d2d5-4ba0-819a-aa0f253e378c" (UID: "156bcac8-d2d5-4ba0-819a-aa0f253e378c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.975909 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "489b62db-1897-4b8b-809a-6d2906775d93" (UID: "489b62db-1897-4b8b-809a-6d2906775d93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.976295 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-config" (OuterVolumeSpecName: "config") pod "489b62db-1897-4b8b-809a-6d2906775d93" (UID: "489b62db-1897-4b8b-809a-6d2906775d93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.982186 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489b62db-1897-4b8b-809a-6d2906775d93-kube-api-access-9bhbw" (OuterVolumeSpecName: "kube-api-access-9bhbw") pod "489b62db-1897-4b8b-809a-6d2906775d93" (UID: "489b62db-1897-4b8b-809a-6d2906775d93"). InnerVolumeSpecName "kube-api-access-9bhbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:12 crc kubenswrapper[4675]: I0219 09:00:12.983129 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156bcac8-d2d5-4ba0-819a-aa0f253e378c-kube-api-access-m45tw" (OuterVolumeSpecName: "kube-api-access-m45tw") pod "156bcac8-d2d5-4ba0-819a-aa0f253e378c" (UID: "156bcac8-d2d5-4ba0-819a-aa0f253e378c"). InnerVolumeSpecName "kube-api-access-m45tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.076540 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156bcac8-d2d5-4ba0-819a-aa0f253e378c-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.076576 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.076587 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bhbw\" (UniqueName: \"kubernetes.io/projected/489b62db-1897-4b8b-809a-6d2906775d93-kube-api-access-9bhbw\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.076599 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m45tw\" (UniqueName: \"kubernetes.io/projected/156bcac8-d2d5-4ba0-819a-aa0f253e378c-kube-api-access-m45tw\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.076607 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/489b62db-1897-4b8b-809a-6d2906775d93-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.132299 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1adde38f-79d6-4ee4-82ed-97004ffb7f85","Type":"ContainerStarted","Data":"a993f5ff21a91f8f3b65ffb590d4332551df482a908b1a402992186078332313"} Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.135133 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" event={"ID":"156bcac8-d2d5-4ba0-819a-aa0f253e378c","Type":"ContainerDied","Data":"f066db28f76aaef45938da64932c5c32c17e1eea421f54fe3d09d90aa78ddd11"} Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.135220 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vzbdl" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.137579 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" event={"ID":"489b62db-1897-4b8b-809a-6d2906775d93","Type":"ContainerDied","Data":"6d248e18472cfeb7d2bb7852d3e1317ddcd3f4e1c7880a984679851bf7113bed"} Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.137704 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-d9gkd" Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.240287 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vzbdl"] Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.248050 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vzbdl"] Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.280761 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d9gkd"] Feb 19 09:00:13 crc kubenswrapper[4675]: I0219 09:00:13.289652 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-d9gkd"] Feb 19 09:00:14 crc kubenswrapper[4675]: I0219 09:00:14.147373 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a68599d-2cc4-4323-8437-58d0866e994e" containerID="c84abb88f7fe6735590038f2294263e3ef8bea42d52051f39cbff23483fd4029" exitCode=0 Feb 19 09:00:14 crc kubenswrapper[4675]: I0219 09:00:14.148751 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" event={"ID":"7a68599d-2cc4-4323-8437-58d0866e994e","Type":"ContainerDied","Data":"c84abb88f7fe6735590038f2294263e3ef8bea42d52051f39cbff23483fd4029"} Feb 19 09:00:15 crc kubenswrapper[4675]: I0219 09:00:15.112541 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156bcac8-d2d5-4ba0-819a-aa0f253e378c" path="/var/lib/kubelet/pods/156bcac8-d2d5-4ba0-819a-aa0f253e378c/volumes" Feb 19 09:00:15 crc kubenswrapper[4675]: I0219 09:00:15.112943 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="489b62db-1897-4b8b-809a-6d2906775d93" path="/var/lib/kubelet/pods/489b62db-1897-4b8b-809a-6d2906775d93/volumes" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.067416 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.170047 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" event={"ID":"7a68599d-2cc4-4323-8437-58d0866e994e","Type":"ContainerDied","Data":"7583ac43950a14c75c4da27823130b38321389535d0104976dfedeee43c00607"} Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.170413 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7583ac43950a14c75c4da27823130b38321389535d0104976dfedeee43c00607" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.170511 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.184854 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a68599d-2cc4-4323-8437-58d0866e994e-config-volume\") pod \"7a68599d-2cc4-4323-8437-58d0866e994e\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.185228 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75kww\" (UniqueName: \"kubernetes.io/projected/7a68599d-2cc4-4323-8437-58d0866e994e-kube-api-access-75kww\") pod \"7a68599d-2cc4-4323-8437-58d0866e994e\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.185378 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a68599d-2cc4-4323-8437-58d0866e994e-secret-volume\") pod \"7a68599d-2cc4-4323-8437-58d0866e994e\" (UID: \"7a68599d-2cc4-4323-8437-58d0866e994e\") " Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.185796 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a68599d-2cc4-4323-8437-58d0866e994e-config-volume" (OuterVolumeSpecName: "config-volume") pod "7a68599d-2cc4-4323-8437-58d0866e994e" (UID: "7a68599d-2cc4-4323-8437-58d0866e994e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.186028 4675 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a68599d-2cc4-4323-8437-58d0866e994e-config-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.278329 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a68599d-2cc4-4323-8437-58d0866e994e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7a68599d-2cc4-4323-8437-58d0866e994e" (UID: "7a68599d-2cc4-4323-8437-58d0866e994e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.278957 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a68599d-2cc4-4323-8437-58d0866e994e-kube-api-access-75kww" (OuterVolumeSpecName: "kube-api-access-75kww") pod "7a68599d-2cc4-4323-8437-58d0866e994e" (UID: "7a68599d-2cc4-4323-8437-58d0866e994e"). InnerVolumeSpecName "kube-api-access-75kww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.287678 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75kww\" (UniqueName: \"kubernetes.io/projected/7a68599d-2cc4-4323-8437-58d0866e994e-kube-api-access-75kww\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:17 crc kubenswrapper[4675]: I0219 09:00:17.287709 4675 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a68599d-2cc4-4323-8437-58d0866e994e-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.177629 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4f51af02-6352-4462-b00f-a2feafe19f23","Type":"ContainerStarted","Data":"67d6993e95dfd3e215db69d7d156642932b80446da1d2b5ce0d133ec5d8a762a"} Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.178600 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.180683 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ba2e218a-edb2-4626-b480-1e4a534bc80d","Type":"ContainerStarted","Data":"7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3"} Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.181213 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.185790 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2344b53c-5df6-4673-beff-7c3cdd28dc93","Type":"ContainerStarted","Data":"fa771f1c34e09a794acbda9783e06b3f61b791f3f8493f0aa90c94000600474c"} Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.187654 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"482efc7c-8e44-4c25-903e-5c8b2cc58d0b","Type":"ContainerStarted","Data":"c48445cf3e84679bb408ef24b35714eb1b994a60aad5477de126421359e65b49"} Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.191325 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc" event={"ID":"0b11fc83-5e56-43d3-a340-80686a5bdf35","Type":"ContainerStarted","Data":"f52ca2f981b6071d31a1704112d11d3ddbf267792539f51197cf90b1145a764c"} Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.191428 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vlnhc" Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.193172 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3b301f22-5f49-490c-899a-f518d8da2174","Type":"ContainerStarted","Data":"7d8f16b7bb999c507acdf3f909a5b3774ca85ddf7e4f2630fc8394cc8983d889"} Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.202744 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=28.374643391 podStartE2EDuration="33.202726372s" podCreationTimestamp="2026-02-19 08:59:45 +0000 UTC" firstStartedPulling="2026-02-19 09:00:12.04378826 +0000 UTC m=+993.670878528" lastFinishedPulling="2026-02-19 09:00:16.871871241 +0000 UTC m=+998.498961509" observedRunningTime="2026-02-19 09:00:18.198268189 +0000 UTC m=+999.825358467" watchObservedRunningTime="2026-02-19 09:00:18.202726372 +0000 UTC m=+999.829816640" Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.241586 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=25.704159511 podStartE2EDuration="31.24156585s" podCreationTimestamp="2026-02-19 08:59:47 +0000 UTC" firstStartedPulling="2026-02-19 09:00:11.331662761 +0000 UTC m=+992.958753029" lastFinishedPulling="2026-02-19 09:00:16.8690691 +0000 UTC m=+998.496159368" observedRunningTime="2026-02-19 09:00:18.235536836 +0000 UTC m=+999.862627104" watchObservedRunningTime="2026-02-19 09:00:18.24156585 +0000 UTC m=+999.868656128" Feb 19 09:00:18 crc kubenswrapper[4675]: I0219 09:00:18.261550 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vlnhc" podStartSLOduration=23.02598859 podStartE2EDuration="28.261533667s" podCreationTimestamp="2026-02-19 08:59:50 +0000 UTC" firstStartedPulling="2026-02-19 09:00:11.656088776 +0000 UTC m=+993.283179044" lastFinishedPulling="2026-02-19 09:00:16.891633853 +0000 UTC m=+998.518724121" observedRunningTime="2026-02-19 09:00:18.255152235 +0000 UTC m=+999.882242503" watchObservedRunningTime="2026-02-19 09:00:18.261533667 +0000 UTC m=+999.888623935" Feb 19 09:00:19 crc kubenswrapper[4675]: I0219 09:00:19.205343 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93dbb123-819d-44f7-80f5-8c38c639f5f4","Type":"ContainerStarted","Data":"897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b"} Feb 19 09:00:19 crc kubenswrapper[4675]: I0219 09:00:19.210579 4675 generic.go:334] "Generic (PLEG): container finished" podID="6dd9438b-4fad-4a66-8d08-df4816a19124" containerID="9fdedb461f1f5b6aaf3067c03e4624e4f2bb633b7e84f145f8242ed55b2d91e3" exitCode=0 Feb 19 09:00:19 crc kubenswrapper[4675]: I0219 09:00:19.210659 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5dht" event={"ID":"6dd9438b-4fad-4a66-8d08-df4816a19124","Type":"ContainerDied","Data":"9fdedb461f1f5b6aaf3067c03e4624e4f2bb633b7e84f145f8242ed55b2d91e3"} Feb 19 09:00:19 crc kubenswrapper[4675]: I0219 09:00:19.219179 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f","Type":"ContainerStarted","Data":"48bce06926e03dcee9ac85de6d1b5b78fdeda47d2d4a80db8727ea95fcd391cc"} Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.227681 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2344b53c-5df6-4673-beff-7c3cdd28dc93","Type":"ContainerStarted","Data":"fd4a5f72f4fe0488e174ec09c35f83880698bba5aefdb5a9c4217b82c086eda5"} Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.229161 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"482efc7c-8e44-4c25-903e-5c8b2cc58d0b","Type":"ContainerStarted","Data":"5276a888df80f2e6a67cf626e4c3b9e37a62a49be98207abef7975df537bbbae"} Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.232007 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5dht" event={"ID":"6dd9438b-4fad-4a66-8d08-df4816a19124","Type":"ContainerStarted","Data":"d259dbf4b94bed1e63263feee94623fc67864fea92eb20333e96d2a1160c93d7"} Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.232030 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5dht" event={"ID":"6dd9438b-4fad-4a66-8d08-df4816a19124","Type":"ContainerStarted","Data":"ad4f6fa1d81f419cc94775d4ff065e6aa47b6abbf526c4743b7ead045d81ae8f"} Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.232042 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-s5dht" Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.232450 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-s5dht" Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.251073 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=19.920517877 podStartE2EDuration="28.251054809s" podCreationTimestamp="2026-02-19 08:59:52 +0000 UTC" firstStartedPulling="2026-02-19 09:00:11.331273542 +0000 UTC m=+992.958363810" lastFinishedPulling="2026-02-19 09:00:19.661810454 +0000 UTC m=+1001.288900742" observedRunningTime="2026-02-19 09:00:20.245255522 +0000 UTC m=+1001.872345790" watchObservedRunningTime="2026-02-19 09:00:20.251054809 +0000 UTC m=+1001.878145077" Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.277122 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=19.684893162 podStartE2EDuration="27.277105221s" podCreationTimestamp="2026-02-19 08:59:53 +0000 UTC" firstStartedPulling="2026-02-19 09:00:12.061451328 +0000 UTC m=+993.688541596" lastFinishedPulling="2026-02-19 09:00:19.653663387 +0000 UTC m=+1001.280753655" observedRunningTime="2026-02-19 09:00:20.27350167 +0000 UTC m=+1001.900591938" watchObservedRunningTime="2026-02-19 09:00:20.277105221 +0000 UTC m=+1001.904195489" Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.302147 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-s5dht" podStartSLOduration=24.782718216 podStartE2EDuration="30.302112667s" podCreationTimestamp="2026-02-19 08:59:50 +0000 UTC" firstStartedPulling="2026-02-19 09:00:11.331934388 +0000 UTC m=+992.959024656" lastFinishedPulling="2026-02-19 09:00:16.851328839 +0000 UTC m=+998.478419107" observedRunningTime="2026-02-19 09:00:20.296932055 +0000 UTC m=+1001.924022353" watchObservedRunningTime="2026-02-19 09:00:20.302112667 +0000 UTC m=+1001.929202935" Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.427786 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 19 09:00:20 crc kubenswrapper[4675]: I0219 09:00:20.465569 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 19 09:00:21 crc kubenswrapper[4675]: I0219 09:00:21.237746 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.224075 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.249454 4675 generic.go:334] "Generic (PLEG): container finished" podID="b91a4d30-10a1-4827-b4a7-2c620a2c2d9f" containerID="48bce06926e03dcee9ac85de6d1b5b78fdeda47d2d4a80db8727ea95fcd391cc" exitCode=0 Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.249539 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f","Type":"ContainerDied","Data":"48bce06926e03dcee9ac85de6d1b5b78fdeda47d2d4a80db8727ea95fcd391cc"} Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.252024 4675 generic.go:334] "Generic (PLEG): container finished" podID="3b301f22-5f49-490c-899a-f518d8da2174" containerID="7d8f16b7bb999c507acdf3f909a5b3774ca85ddf7e4f2630fc8394cc8983d889" exitCode=0 Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.252189 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3b301f22-5f49-490c-899a-f518d8da2174","Type":"ContainerDied","Data":"7d8f16b7bb999c507acdf3f909a5b3774ca85ddf7e4f2630fc8394cc8983d889"} Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.283780 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.350103 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.606943 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chnjb"] Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.666843 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6w7l7"] Feb 19 09:00:22 crc kubenswrapper[4675]: E0219 09:00:22.667211 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a68599d-2cc4-4323-8437-58d0866e994e" containerName="collect-profiles" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.667229 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a68599d-2cc4-4323-8437-58d0866e994e" containerName="collect-profiles" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.667372 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a68599d-2cc4-4323-8437-58d0866e994e" containerName="collect-profiles" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.668250 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.671922 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.678950 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6w7l7"] Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.703415 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lnndq"] Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.704739 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.707825 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.749305 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lnndq"] Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774268 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dg4q\" (UniqueName: \"kubernetes.io/projected/10a37557-1329-4995-86a8-4dc2e12f217c-kube-api-access-2dg4q\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774306 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a37557-1329-4995-86a8-4dc2e12f217c-config\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774329 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a37557-1329-4995-86a8-4dc2e12f217c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774366 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47nhv\" (UniqueName: \"kubernetes.io/projected/2aca9be2-1569-47b8-bbaa-ecd11352ef52-kube-api-access-47nhv\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774417 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/10a37557-1329-4995-86a8-4dc2e12f217c-ovn-rundir\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774461 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/10a37557-1329-4995-86a8-4dc2e12f217c-ovs-rundir\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774492 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774513 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774535 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-config\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.774553 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a37557-1329-4995-86a8-4dc2e12f217c-combined-ca-bundle\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.875525 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/10a37557-1329-4995-86a8-4dc2e12f217c-ovn-rundir\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876000 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/10a37557-1329-4995-86a8-4dc2e12f217c-ovs-rundir\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876039 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876068 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876073 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/10a37557-1329-4995-86a8-4dc2e12f217c-ovs-rundir\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876096 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-config\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876220 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a37557-1329-4995-86a8-4dc2e12f217c-combined-ca-bundle\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876305 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dg4q\" (UniqueName: \"kubernetes.io/projected/10a37557-1329-4995-86a8-4dc2e12f217c-kube-api-access-2dg4q\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876323 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a37557-1329-4995-86a8-4dc2e12f217c-config\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876351 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a37557-1329-4995-86a8-4dc2e12f217c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.876394 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47nhv\" (UniqueName: \"kubernetes.io/projected/2aca9be2-1569-47b8-bbaa-ecd11352ef52-kube-api-access-47nhv\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.875988 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/10a37557-1329-4995-86a8-4dc2e12f217c-ovn-rundir\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.878275 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.878477 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-config\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.878868 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.881157 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a37557-1329-4995-86a8-4dc2e12f217c-config\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.896963 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a37557-1329-4995-86a8-4dc2e12f217c-combined-ca-bundle\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.899462 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a37557-1329-4995-86a8-4dc2e12f217c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.903602 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dg4q\" (UniqueName: \"kubernetes.io/projected/10a37557-1329-4995-86a8-4dc2e12f217c-kube-api-access-2dg4q\") pod \"ovn-controller-metrics-lnndq\" (UID: \"10a37557-1329-4995-86a8-4dc2e12f217c\") " pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.905882 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47nhv\" (UniqueName: \"kubernetes.io/projected/2aca9be2-1569-47b8-bbaa-ecd11352ef52-kube-api-access-47nhv\") pod \"dnsmasq-dns-6bc7876d45-6w7l7\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.956850 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 09:00:22 crc kubenswrapper[4675]: I0219 09:00:22.988665 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.032110 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lnndq" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.038948 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nnzpk"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.076594 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-s62wm"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.080498 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.081754 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7jb7\" (UniqueName: \"kubernetes.io/projected/a5a4aa24-a744-45fb-88a6-546affb72d83-kube-api-access-z7jb7\") pod \"a5a4aa24-a744-45fb-88a6-546affb72d83\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.081817 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-config\") pod \"a5a4aa24-a744-45fb-88a6-546affb72d83\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.081897 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-dns-svc\") pod \"a5a4aa24-a744-45fb-88a6-546affb72d83\" (UID: \"a5a4aa24-a744-45fb-88a6-546affb72d83\") " Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.082630 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a5a4aa24-a744-45fb-88a6-546affb72d83" (UID: "a5a4aa24-a744-45fb-88a6-546affb72d83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.083020 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-config" (OuterVolumeSpecName: "config") pod "a5a4aa24-a744-45fb-88a6-546affb72d83" (UID: "a5a4aa24-a744-45fb-88a6-546affb72d83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.083185 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.088479 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-s62wm"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.090063 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a4aa24-a744-45fb-88a6-546affb72d83-kube-api-access-z7jb7" (OuterVolumeSpecName: "kube-api-access-z7jb7") pod "a5a4aa24-a744-45fb-88a6-546affb72d83" (UID: "a5a4aa24-a744-45fb-88a6-546affb72d83"). InnerVolumeSpecName "kube-api-access-z7jb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.184735 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.184782 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-dns-svc\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.184855 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-config\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.184896 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2c5b\" (UniqueName: \"kubernetes.io/projected/656cfb75-4638-4306-8490-1854667f58d0-kube-api-access-t2c5b\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.184976 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.185056 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7jb7\" (UniqueName: \"kubernetes.io/projected/a5a4aa24-a744-45fb-88a6-546affb72d83-kube-api-access-z7jb7\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.185067 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.185097 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5a4aa24-a744-45fb-88a6-546affb72d83-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.263261 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" event={"ID":"a5a4aa24-a744-45fb-88a6-546affb72d83","Type":"ContainerDied","Data":"3d5074c9c648dac25d6f24ca87868c6df8b10a78d999d5418062c87e5cc73041"} Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.270591 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b91a4d30-10a1-4827-b4a7-2c620a2c2d9f","Type":"ContainerStarted","Data":"d1d5a4ca8bf2bc33cfebed042d19968ad5f526a804943fa017266b8ced59630e"} Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.275542 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chnjb" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.276247 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3b301f22-5f49-490c-899a-f518d8da2174","Type":"ContainerStarted","Data":"f085f192747bba02895383e00c183bf9341a6b82e449080142bfa226e1733dfa"} Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.276304 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.286717 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.287130 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.287153 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-dns-svc\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.287183 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-config\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.287220 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2c5b\" (UniqueName: \"kubernetes.io/projected/656cfb75-4638-4306-8490-1854667f58d0-kube-api-access-t2c5b\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.289012 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.290926 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.292138 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-dns-svc\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.292600 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371995.562191 podStartE2EDuration="41.292584947s" podCreationTimestamp="2026-02-19 08:59:42 +0000 UTC" firstStartedPulling="2026-02-19 08:59:44.765301847 +0000 UTC m=+966.392392115" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:00:23.292429063 +0000 UTC m=+1004.919519331" watchObservedRunningTime="2026-02-19 09:00:23.292584947 +0000 UTC m=+1004.919675215" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.292913 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-config\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.319404 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2c5b\" (UniqueName: \"kubernetes.io/projected/656cfb75-4638-4306-8490-1854667f58d0-kube-api-access-t2c5b\") pod \"dnsmasq-dns-8554648995-s62wm\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.324109 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=35.996509701 podStartE2EDuration="39.324088788s" podCreationTimestamp="2026-02-19 08:59:44 +0000 UTC" firstStartedPulling="2026-02-19 09:00:12.065121062 +0000 UTC m=+993.692211330" lastFinishedPulling="2026-02-19 09:00:15.392700149 +0000 UTC m=+997.019790417" observedRunningTime="2026-02-19 09:00:23.323657957 +0000 UTC m=+1004.950748225" watchObservedRunningTime="2026-02-19 09:00:23.324088788 +0000 UTC m=+1004.951179056" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.366959 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chnjb"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.372733 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chnjb"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.375153 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.407860 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.548817 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.558150 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.574911 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.577043 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.587186 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.592183 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.592446 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-cj554" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.599230 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.611841 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-dns-svc\") pod \"dc6d527a-29ec-432a-a7c1-87a54af36db8\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.612114 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmf9n\" (UniqueName: \"kubernetes.io/projected/dc6d527a-29ec-432a-a7c1-87a54af36db8-kube-api-access-wmf9n\") pod \"dc6d527a-29ec-432a-a7c1-87a54af36db8\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.612330 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-config\") pod \"dc6d527a-29ec-432a-a7c1-87a54af36db8\" (UID: \"dc6d527a-29ec-432a-a7c1-87a54af36db8\") " Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.612714 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.612852 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9d70128-744b-4075-ab12-f60f6c1838e7-scripts\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.613053 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.613152 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d9d70128-744b-4075-ab12-f60f6c1838e7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.613282 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d70128-744b-4075-ab12-f60f6c1838e7-config\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.613393 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.613537 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9qtf\" (UniqueName: \"kubernetes.io/projected/d9d70128-744b-4075-ab12-f60f6c1838e7-kube-api-access-t9qtf\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.614463 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc6d527a-29ec-432a-a7c1-87a54af36db8" (UID: "dc6d527a-29ec-432a-a7c1-87a54af36db8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.619805 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-config" (OuterVolumeSpecName: "config") pod "dc6d527a-29ec-432a-a7c1-87a54af36db8" (UID: "dc6d527a-29ec-432a-a7c1-87a54af36db8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.626212 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc6d527a-29ec-432a-a7c1-87a54af36db8-kube-api-access-wmf9n" (OuterVolumeSpecName: "kube-api-access-wmf9n") pod "dc6d527a-29ec-432a-a7c1-87a54af36db8" (UID: "dc6d527a-29ec-432a-a7c1-87a54af36db8"). InnerVolumeSpecName "kube-api-access-wmf9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.667260 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6w7l7"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.699955 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lnndq"] Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.714756 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d9d70128-744b-4075-ab12-f60f6c1838e7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.714831 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d70128-744b-4075-ab12-f60f6c1838e7-config\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.714873 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.714928 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9qtf\" (UniqueName: \"kubernetes.io/projected/d9d70128-744b-4075-ab12-f60f6c1838e7-kube-api-access-t9qtf\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.714984 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.715270 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9d70128-744b-4075-ab12-f60f6c1838e7-scripts\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.715324 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.715381 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.715397 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmf9n\" (UniqueName: \"kubernetes.io/projected/dc6d527a-29ec-432a-a7c1-87a54af36db8-kube-api-access-wmf9n\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.715412 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc6d527a-29ec-432a-a7c1-87a54af36db8-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.715929 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9d70128-744b-4075-ab12-f60f6c1838e7-config\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.716189 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d9d70128-744b-4075-ab12-f60f6c1838e7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.716746 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9d70128-744b-4075-ab12-f60f6c1838e7-scripts\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.719036 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: W0219 09:00:23.719085 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a37557_1329_4995_86a8_4dc2e12f217c.slice/crio-61aa3c642910949a108a56968c958c2ece42133c39157b16c6fd256c0ea11391 WatchSource:0}: Error finding container 61aa3c642910949a108a56968c958c2ece42133c39157b16c6fd256c0ea11391: Status 404 returned error can't find the container with id 61aa3c642910949a108a56968c958c2ece42133c39157b16c6fd256c0ea11391 Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.719556 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.720698 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d70128-744b-4075-ab12-f60f6c1838e7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.743693 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9qtf\" (UniqueName: \"kubernetes.io/projected/d9d70128-744b-4075-ab12-f60f6c1838e7-kube-api-access-t9qtf\") pod \"ovn-northd-0\" (UID: \"d9d70128-744b-4075-ab12-f60f6c1838e7\") " pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.917007 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 19 09:00:23 crc kubenswrapper[4675]: I0219 09:00:23.999336 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-s62wm"] Feb 19 09:00:24 crc kubenswrapper[4675]: W0219 09:00:24.007739 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod656cfb75_4638_4306_8490_1854667f58d0.slice/crio-5b838a9fb800e92c497801db6d9a7fc09cbc67ffc614d5fdf7a3be2031ab8b22 WatchSource:0}: Error finding container 5b838a9fb800e92c497801db6d9a7fc09cbc67ffc614d5fdf7a3be2031ab8b22: Status 404 returned error can't find the container with id 5b838a9fb800e92c497801db6d9a7fc09cbc67ffc614d5fdf7a3be2031ab8b22 Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.119286 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.119321 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.186045 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 19 09:00:24 crc kubenswrapper[4675]: W0219 09:00:24.193783 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9d70128_744b_4075_ab12_f60f6c1838e7.slice/crio-0ac7cf44109d478cdc46d093626d3ea3cc566257c377928625736353446c8046 WatchSource:0}: Error finding container 0ac7cf44109d478cdc46d093626d3ea3cc566257c377928625736353446c8046: Status 404 returned error can't find the container with id 0ac7cf44109d478cdc46d093626d3ea3cc566257c377928625736353446c8046 Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.288241 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d9d70128-744b-4075-ab12-f60f6c1838e7","Type":"ContainerStarted","Data":"0ac7cf44109d478cdc46d093626d3ea3cc566257c377928625736353446c8046"} Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.289238 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" event={"ID":"2aca9be2-1569-47b8-bbaa-ecd11352ef52","Type":"ContainerStarted","Data":"0263344c5abfcff5e5cf781e307175b6a85fcd56414e8bd6db03ae5ccfd77624"} Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.290460 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lnndq" event={"ID":"10a37557-1329-4995-86a8-4dc2e12f217c","Type":"ContainerStarted","Data":"28c076a515f84d9ac14b2ad8f3e48089765abfcffe335767cd0415995d52d8d8"} Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.290530 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lnndq" event={"ID":"10a37557-1329-4995-86a8-4dc2e12f217c","Type":"ContainerStarted","Data":"61aa3c642910949a108a56968c958c2ece42133c39157b16c6fd256c0ea11391"} Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.291358 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-s62wm" event={"ID":"656cfb75-4638-4306-8490-1854667f58d0","Type":"ContainerStarted","Data":"5b838a9fb800e92c497801db6d9a7fc09cbc67ffc614d5fdf7a3be2031ab8b22"} Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.295824 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.309596 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nnzpk" event={"ID":"dc6d527a-29ec-432a-a7c1-87a54af36db8","Type":"ContainerDied","Data":"f8334309e751429de474f4a441cf69cdcafcd463ab19354d509a0c4e3868b5e6"} Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.346147 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lnndq" podStartSLOduration=2.346129432 podStartE2EDuration="2.346129432s" podCreationTimestamp="2026-02-19 09:00:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:00:24.340201591 +0000 UTC m=+1005.967291869" watchObservedRunningTime="2026-02-19 09:00:24.346129432 +0000 UTC m=+1005.973219700" Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.395013 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nnzpk"] Feb 19 09:00:24 crc kubenswrapper[4675]: I0219 09:00:24.410268 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nnzpk"] Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.113838 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a4aa24-a744-45fb-88a6-546affb72d83" path="/var/lib/kubelet/pods/a5a4aa24-a744-45fb-88a6-546affb72d83/volumes" Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.114454 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc6d527a-29ec-432a-a7c1-87a54af36db8" path="/var/lib/kubelet/pods/dc6d527a-29ec-432a-a7c1-87a54af36db8/volumes" Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.310407 4675 generic.go:334] "Generic (PLEG): container finished" podID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerID="332d7e669c40c26f1b64d19542c71d2711464e3fe1e513eff59810913c25984e" exitCode=0 Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.310473 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" event={"ID":"2aca9be2-1569-47b8-bbaa-ecd11352ef52","Type":"ContainerDied","Data":"332d7e669c40c26f1b64d19542c71d2711464e3fe1e513eff59810913c25984e"} Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.314457 4675 generic.go:334] "Generic (PLEG): container finished" podID="656cfb75-4638-4306-8490-1854667f58d0" containerID="9f6e34c1494f033243540ebd9f9c3d53814851da486fd1c429ea5292b803b6c1" exitCode=0 Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.315273 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-s62wm" event={"ID":"656cfb75-4638-4306-8490-1854667f58d0","Type":"ContainerDied","Data":"9f6e34c1494f033243540ebd9f9c3d53814851da486fd1c429ea5292b803b6c1"} Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.551436 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.552032 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 19 09:00:25 crc kubenswrapper[4675]: I0219 09:00:25.835379 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.322216 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d9d70128-744b-4075-ab12-f60f6c1838e7","Type":"ContainerStarted","Data":"9516342c207819642adb61ca19d25f241133796f6db70fdc6578e72f8326cef7"} Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.322266 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d9d70128-744b-4075-ab12-f60f6c1838e7","Type":"ContainerStarted","Data":"1e7ac6ca1a8f8f77fcaac66a5e3c9849bbe59c1e33cefa4d06122bdc00b5b931"} Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.322309 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.323753 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" event={"ID":"2aca9be2-1569-47b8-bbaa-ecd11352ef52","Type":"ContainerStarted","Data":"3b0a5cb6f1678e27d5140e7bce76345bbc752a45de6d066d69161efd52b5a75d"} Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.323874 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.325329 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-s62wm" event={"ID":"656cfb75-4638-4306-8490-1854667f58d0","Type":"ContainerStarted","Data":"f6a6f3613036a1b2e8ed89e7479a9d36466e291c3e396940458f3b37fee9324e"} Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.342463 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.242522913 podStartE2EDuration="3.342446106s" podCreationTimestamp="2026-02-19 09:00:23 +0000 UTC" firstStartedPulling="2026-02-19 09:00:24.195988546 +0000 UTC m=+1005.823078824" lastFinishedPulling="2026-02-19 09:00:25.295911749 +0000 UTC m=+1006.923002017" observedRunningTime="2026-02-19 09:00:26.339094811 +0000 UTC m=+1007.966185079" watchObservedRunningTime="2026-02-19 09:00:26.342446106 +0000 UTC m=+1007.969536374" Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.360858 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" podStartSLOduration=3.9165355330000002 podStartE2EDuration="4.360838404s" podCreationTimestamp="2026-02-19 09:00:22 +0000 UTC" firstStartedPulling="2026-02-19 09:00:23.676285269 +0000 UTC m=+1005.303375547" lastFinishedPulling="2026-02-19 09:00:24.12058815 +0000 UTC m=+1005.747678418" observedRunningTime="2026-02-19 09:00:26.355244011 +0000 UTC m=+1007.982334279" watchObservedRunningTime="2026-02-19 09:00:26.360838404 +0000 UTC m=+1007.987928672" Feb 19 09:00:26 crc kubenswrapper[4675]: I0219 09:00:26.375848 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-s62wm" podStartSLOduration=2.745616009 podStartE2EDuration="3.375828885s" podCreationTimestamp="2026-02-19 09:00:23 +0000 UTC" firstStartedPulling="2026-02-19 09:00:24.012172785 +0000 UTC m=+1005.639263053" lastFinishedPulling="2026-02-19 09:00:24.642385661 +0000 UTC m=+1006.269475929" observedRunningTime="2026-02-19 09:00:26.37208827 +0000 UTC m=+1007.999178528" watchObservedRunningTime="2026-02-19 09:00:26.375828885 +0000 UTC m=+1008.002919153" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.333384 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.889195 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6w7l7"] Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.924204 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4b59q"] Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.927743 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.947313 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4b59q"] Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.991575 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.997198 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.997241 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqstf\" (UniqueName: \"kubernetes.io/projected/ade75766-b387-4393-b66b-5a09665137fb-kube-api-access-lqstf\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.997276 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.997516 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:27 crc kubenswrapper[4675]: I0219 09:00:27.997622 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-config\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.048972 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.099762 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.099832 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-config\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.099935 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.099961 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqstf\" (UniqueName: \"kubernetes.io/projected/ade75766-b387-4393-b66b-5a09665137fb-kube-api-access-lqstf\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.099987 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.100955 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.100965 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.101008 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.101225 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-config\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.120574 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqstf\" (UniqueName: \"kubernetes.io/projected/ade75766-b387-4393-b66b-5a09665137fb-kube-api-access-lqstf\") pod \"dnsmasq-dns-b8fbc5445-4b59q\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.201611 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.250251 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.347004 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" podUID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerName="dnsmasq-dns" containerID="cri-o://3b0a5cb6f1678e27d5140e7bce76345bbc752a45de6d066d69161efd52b5a75d" gracePeriod=10 Feb 19 09:00:28 crc kubenswrapper[4675]: I0219 09:00:28.815583 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4b59q"] Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.042976 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.053983 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.056276 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.056936 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-twc2v" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.056986 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.060603 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.068752 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.139656 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.139708 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.139747 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-lock\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.139791 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrcc2\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-kube-api-access-vrcc2\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.139920 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-cache\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.139944 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.241892 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.242288 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.242336 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-lock\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.242374 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrcc2\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-kube-api-access-vrcc2\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.242394 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-cache\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.242412 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: E0219 09:00:29.242563 4675 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 19 09:00:29 crc kubenswrapper[4675]: E0219 09:00:29.242578 4675 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 19 09:00:29 crc kubenswrapper[4675]: E0219 09:00:29.242623 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift podName:74a1ef5b-9639-489e-9dc7-2da01bffa6f2 nodeName:}" failed. No retries permitted until 2026-02-19 09:00:29.742608881 +0000 UTC m=+1011.369699149 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift") pod "swift-storage-0" (UID: "74a1ef5b-9639-489e-9dc7-2da01bffa6f2") : configmap "swift-ring-files" not found Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.242733 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.243038 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-lock\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.243106 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-cache\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.247842 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.260685 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrcc2\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-kube-api-access-vrcc2\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.272428 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.349857 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" event={"ID":"ade75766-b387-4393-b66b-5a09665137fb","Type":"ContainerStarted","Data":"a45d6e31321f8650e4268d01d4cc5bd18c6747e45c0e17975231f18a1d306bcb"} Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.352161 4675 generic.go:334] "Generic (PLEG): container finished" podID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerID="3b0a5cb6f1678e27d5140e7bce76345bbc752a45de6d066d69161efd52b5a75d" exitCode=0 Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.352192 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" event={"ID":"2aca9be2-1569-47b8-bbaa-ecd11352ef52","Type":"ContainerDied","Data":"3b0a5cb6f1678e27d5140e7bce76345bbc752a45de6d066d69161efd52b5a75d"} Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.666464 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wkwcx"] Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.668072 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.671872 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.671873 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.671971 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.694975 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wkwcx"] Feb 19 09:00:29 crc kubenswrapper[4675]: E0219 09:00:29.695532 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-4p8zw ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-4p8zw ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-wkwcx" podUID="9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.710955 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wkwcx"] Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.715954 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-s4x6m"] Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.716967 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.733965 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-s4x6m"] Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.751517 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-etc-swift\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.751896 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p8zw\" (UniqueName: \"kubernetes.io/projected/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-kube-api-access-4p8zw\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.752007 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-scripts\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.752170 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.752294 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-dispersionconf\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.752398 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-ring-data-devices\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.752597 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-combined-ca-bundle\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.752770 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-swiftconf\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: E0219 09:00:29.752980 4675 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 19 09:00:29 crc kubenswrapper[4675]: E0219 09:00:29.753107 4675 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 19 09:00:29 crc kubenswrapper[4675]: E0219 09:00:29.753226 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift podName:74a1ef5b-9639-489e-9dc7-2da01bffa6f2 nodeName:}" failed. No retries permitted until 2026-02-19 09:00:30.753210727 +0000 UTC m=+1012.380300995 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift") pod "swift-storage-0" (UID: "74a1ef5b-9639-489e-9dc7-2da01bffa6f2") : configmap "swift-ring-files" not found Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855556 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-scripts\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855626 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-swiftconf\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855671 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-dispersionconf\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855712 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-ring-data-devices\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855746 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-scripts\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855765 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-combined-ca-bundle\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855814 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dadf56d2-1dd1-4590-8fd7-0cc266af5377-etc-swift\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855844 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-swiftconf\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855868 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-ring-data-devices\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855888 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jqgc\" (UniqueName: \"kubernetes.io/projected/dadf56d2-1dd1-4590-8fd7-0cc266af5377-kube-api-access-6jqgc\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855916 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-etc-swift\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855950 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-combined-ca-bundle\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855966 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-dispersionconf\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.855985 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p8zw\" (UniqueName: \"kubernetes.io/projected/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-kube-api-access-4p8zw\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.857208 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-scripts\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.857561 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-etc-swift\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.857662 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-ring-data-devices\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.866525 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-combined-ca-bundle\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.867026 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-dispersionconf\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.871063 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-swiftconf\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.907437 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p8zw\" (UniqueName: \"kubernetes.io/projected/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-kube-api-access-4p8zw\") pod \"swift-ring-rebalance-wkwcx\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.959690 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-combined-ca-bundle\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.959753 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-dispersionconf\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.959815 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-swiftconf\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.959892 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-scripts\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.959935 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dadf56d2-1dd1-4590-8fd7-0cc266af5377-etc-swift\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.959980 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-ring-data-devices\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.960001 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jqgc\" (UniqueName: \"kubernetes.io/projected/dadf56d2-1dd1-4590-8fd7-0cc266af5377-kube-api-access-6jqgc\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.962458 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-scripts\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.962587 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dadf56d2-1dd1-4590-8fd7-0cc266af5377-etc-swift\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.963154 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-ring-data-devices\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.970424 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-combined-ca-bundle\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.971027 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-dispersionconf\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.977072 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-swiftconf\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:29 crc kubenswrapper[4675]: I0219 09:00:29.991331 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jqgc\" (UniqueName: \"kubernetes.io/projected/dadf56d2-1dd1-4590-8fd7-0cc266af5377-kube-api-access-6jqgc\") pod \"swift-ring-rebalance-s4x6m\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.041102 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.231122 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.357797 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.362597 4675 generic.go:334] "Generic (PLEG): container finished" podID="ade75766-b387-4393-b66b-5a09665137fb" containerID="2d0218220dd3aef3e157051f3dabc4234121c059b6dc78682887daed25d83580" exitCode=0 Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.362687 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" event={"ID":"ade75766-b387-4393-b66b-5a09665137fb","Type":"ContainerDied","Data":"2d0218220dd3aef3e157051f3dabc4234121c059b6dc78682887daed25d83580"} Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.364992 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.365091 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.365171 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6w7l7" event={"ID":"2aca9be2-1569-47b8-bbaa-ecd11352ef52","Type":"ContainerDied","Data":"0263344c5abfcff5e5cf781e307175b6a85fcd56414e8bd6db03ae5ccfd77624"} Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.365226 4675 scope.go:117] "RemoveContainer" containerID="3b0a5cb6f1678e27d5140e7bce76345bbc752a45de6d066d69161efd52b5a75d" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.404753 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.465102 4675 scope.go:117] "RemoveContainer" containerID="332d7e669c40c26f1b64d19542c71d2711464e3fe1e513eff59810913c25984e" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.468895 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-config\") pod \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.469514 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-dns-svc\") pod \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.469577 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-ovsdbserver-sb\") pod \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.469609 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47nhv\" (UniqueName: \"kubernetes.io/projected/2aca9be2-1569-47b8-bbaa-ecd11352ef52-kube-api-access-47nhv\") pod \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\" (UID: \"2aca9be2-1569-47b8-bbaa-ecd11352ef52\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.471838 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.475297 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aca9be2-1569-47b8-bbaa-ecd11352ef52-kube-api-access-47nhv" (OuterVolumeSpecName: "kube-api-access-47nhv") pod "2aca9be2-1569-47b8-bbaa-ecd11352ef52" (UID: "2aca9be2-1569-47b8-bbaa-ecd11352ef52"). InnerVolumeSpecName "kube-api-access-47nhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.515083 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-config" (OuterVolumeSpecName: "config") pod "2aca9be2-1569-47b8-bbaa-ecd11352ef52" (UID: "2aca9be2-1569-47b8-bbaa-ecd11352ef52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.520791 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2aca9be2-1569-47b8-bbaa-ecd11352ef52" (UID: "2aca9be2-1569-47b8-bbaa-ecd11352ef52"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.521293 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2aca9be2-1569-47b8-bbaa-ecd11352ef52" (UID: "2aca9be2-1569-47b8-bbaa-ecd11352ef52"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571441 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-scripts\") pod \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571519 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-etc-swift\") pod \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571562 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p8zw\" (UniqueName: \"kubernetes.io/projected/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-kube-api-access-4p8zw\") pod \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571583 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-ring-data-devices\") pod \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571728 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-dispersionconf\") pod \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571749 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-swiftconf\") pod \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571767 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-combined-ca-bundle\") pod \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\" (UID: \"9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b\") " Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.571988 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-scripts" (OuterVolumeSpecName: "scripts") pod "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" (UID: "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.572432 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.572453 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.572467 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47nhv\" (UniqueName: \"kubernetes.io/projected/2aca9be2-1569-47b8-bbaa-ecd11352ef52-kube-api-access-47nhv\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.572479 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.572488 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2aca9be2-1569-47b8-bbaa-ecd11352ef52-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.572764 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" (UID: "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.572858 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" (UID: "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.575330 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-s4x6m"] Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.577732 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" (UID: "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.578492 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-kube-api-access-4p8zw" (OuterVolumeSpecName: "kube-api-access-4p8zw") pod "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" (UID: "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b"). InnerVolumeSpecName "kube-api-access-4p8zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.579692 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" (UID: "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: W0219 09:00:30.583312 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddadf56d2_1dd1_4590_8fd7_0cc266af5377.slice/crio-d358b5269eb79e735682987356149ac77d413989bdfde11915a906153765bb7d WatchSource:0}: Error finding container d358b5269eb79e735682987356149ac77d413989bdfde11915a906153765bb7d: Status 404 returned error can't find the container with id d358b5269eb79e735682987356149ac77d413989bdfde11915a906153765bb7d Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.583449 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" (UID: "9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.674279 4675 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.674755 4675 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.674772 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.674786 4675 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.674800 4675 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.674813 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p8zw\" (UniqueName: \"kubernetes.io/projected/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b-kube-api-access-4p8zw\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.696541 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6w7l7"] Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.702484 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6w7l7"] Feb 19 09:00:30 crc kubenswrapper[4675]: I0219 09:00:30.776467 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:30 crc kubenswrapper[4675]: E0219 09:00:30.776710 4675 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 19 09:00:30 crc kubenswrapper[4675]: E0219 09:00:30.776736 4675 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 19 09:00:30 crc kubenswrapper[4675]: E0219 09:00:30.776797 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift podName:74a1ef5b-9639-489e-9dc7-2da01bffa6f2 nodeName:}" failed. No retries permitted until 2026-02-19 09:00:32.776778451 +0000 UTC m=+1014.403868719 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift") pod "swift-storage-0" (UID: "74a1ef5b-9639-489e-9dc7-2da01bffa6f2") : configmap "swift-ring-files" not found Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.112162 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" path="/var/lib/kubelet/pods/2aca9be2-1569-47b8-bbaa-ecd11352ef52/volumes" Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.376062 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" event={"ID":"ade75766-b387-4393-b66b-5a09665137fb","Type":"ContainerStarted","Data":"f452536d0b73e2b61a8adf745b19b101d7c5a43a6d7d8adb0eec3a4df9e9bdab"} Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.376308 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.379561 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wkwcx" Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.379614 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s4x6m" event={"ID":"dadf56d2-1dd1-4590-8fd7-0cc266af5377","Type":"ContainerStarted","Data":"d358b5269eb79e735682987356149ac77d413989bdfde11915a906153765bb7d"} Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.433132 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" podStartSLOduration=4.433110961 podStartE2EDuration="4.433110961s" podCreationTimestamp="2026-02-19 09:00:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:00:31.402692418 +0000 UTC m=+1013.029782686" watchObservedRunningTime="2026-02-19 09:00:31.433110961 +0000 UTC m=+1013.060201229" Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.456461 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wkwcx"] Feb 19 09:00:31 crc kubenswrapper[4675]: I0219 09:00:31.478673 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-wkwcx"] Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.811664 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:32 crc kubenswrapper[4675]: E0219 09:00:32.811975 4675 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 19 09:00:32 crc kubenswrapper[4675]: E0219 09:00:32.811997 4675 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 19 09:00:32 crc kubenswrapper[4675]: E0219 09:00:32.812060 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift podName:74a1ef5b-9639-489e-9dc7-2da01bffa6f2 nodeName:}" failed. No retries permitted until 2026-02-19 09:00:36.812042824 +0000 UTC m=+1018.439133112 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift") pod "swift-storage-0" (UID: "74a1ef5b-9639-489e-9dc7-2da01bffa6f2") : configmap "swift-ring-files" not found Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.850712 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-6lkh9"] Feb 19 09:00:32 crc kubenswrapper[4675]: E0219 09:00:32.851124 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerName="init" Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.851145 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerName="init" Feb 19 09:00:32 crc kubenswrapper[4675]: E0219 09:00:32.851168 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerName="dnsmasq-dns" Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.851178 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerName="dnsmasq-dns" Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.851401 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aca9be2-1569-47b8-bbaa-ecd11352ef52" containerName="dnsmasq-dns" Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.854086 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.859382 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.863260 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-6lkh9"] Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.913495 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-operator-scripts\") pod \"root-account-create-update-6lkh9\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:32 crc kubenswrapper[4675]: I0219 09:00:32.913761 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w92n\" (UniqueName: \"kubernetes.io/projected/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-kube-api-access-7w92n\") pod \"root-account-create-update-6lkh9\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:33 crc kubenswrapper[4675]: I0219 09:00:33.015490 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-operator-scripts\") pod \"root-account-create-update-6lkh9\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:33 crc kubenswrapper[4675]: I0219 09:00:33.015610 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w92n\" (UniqueName: \"kubernetes.io/projected/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-kube-api-access-7w92n\") pod \"root-account-create-update-6lkh9\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:33 crc kubenswrapper[4675]: I0219 09:00:33.016234 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-operator-scripts\") pod \"root-account-create-update-6lkh9\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:33 crc kubenswrapper[4675]: I0219 09:00:33.067805 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w92n\" (UniqueName: \"kubernetes.io/projected/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-kube-api-access-7w92n\") pod \"root-account-create-update-6lkh9\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:33 crc kubenswrapper[4675]: I0219 09:00:33.118106 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b" path="/var/lib/kubelet/pods/9bce5608-7bb5-4cc2-8550-c26d5ac3dc6b/volumes" Feb 19 09:00:33 crc kubenswrapper[4675]: I0219 09:00:33.174212 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:33 crc kubenswrapper[4675]: I0219 09:00:33.409850 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:34 crc kubenswrapper[4675]: I0219 09:00:34.404511 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s4x6m" event={"ID":"dadf56d2-1dd1-4590-8fd7-0cc266af5377","Type":"ContainerStarted","Data":"f3b6b52a893e965134b4ccedd11807dfd49efb1f4e73f10e2cc0ec21ae2c2b40"} Feb 19 09:00:34 crc kubenswrapper[4675]: I0219 09:00:34.422163 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-s4x6m" podStartSLOduration=1.862606362 podStartE2EDuration="5.422139154s" podCreationTimestamp="2026-02-19 09:00:29 +0000 UTC" firstStartedPulling="2026-02-19 09:00:30.585247703 +0000 UTC m=+1012.212337971" lastFinishedPulling="2026-02-19 09:00:34.144780495 +0000 UTC m=+1015.771870763" observedRunningTime="2026-02-19 09:00:34.419803485 +0000 UTC m=+1016.046893763" watchObservedRunningTime="2026-02-19 09:00:34.422139154 +0000 UTC m=+1016.049229422" Feb 19 09:00:34 crc kubenswrapper[4675]: I0219 09:00:34.520215 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-6lkh9"] Feb 19 09:00:34 crc kubenswrapper[4675]: W0219 09:00:34.523362 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda010b2ee_7a95_45de_aa98_4898ea6cf9ec.slice/crio-95470785b4046bc386d1df97b0cf52b64a80b5e80549c17bfdc51fbc28f3b157 WatchSource:0}: Error finding container 95470785b4046bc386d1df97b0cf52b64a80b5e80549c17bfdc51fbc28f3b157: Status 404 returned error can't find the container with id 95470785b4046bc386d1df97b0cf52b64a80b5e80549c17bfdc51fbc28f3b157 Feb 19 09:00:35 crc kubenswrapper[4675]: I0219 09:00:35.412848 4675 generic.go:334] "Generic (PLEG): container finished" podID="a010b2ee-7a95-45de-aa98-4898ea6cf9ec" containerID="69153e694360781dae004fbfd2eeebd54afe3db2ba964ad120a8cabb0f562602" exitCode=0 Feb 19 09:00:35 crc kubenswrapper[4675]: I0219 09:00:35.413821 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-6lkh9" event={"ID":"a010b2ee-7a95-45de-aa98-4898ea6cf9ec","Type":"ContainerDied","Data":"69153e694360781dae004fbfd2eeebd54afe3db2ba964ad120a8cabb0f562602"} Feb 19 09:00:35 crc kubenswrapper[4675]: I0219 09:00:35.413882 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-6lkh9" event={"ID":"a010b2ee-7a95-45de-aa98-4898ea6cf9ec","Type":"ContainerStarted","Data":"95470785b4046bc386d1df97b0cf52b64a80b5e80549c17bfdc51fbc28f3b157"} Feb 19 09:00:35 crc kubenswrapper[4675]: I0219 09:00:35.988498 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-hl8ff"] Feb 19 09:00:35 crc kubenswrapper[4675]: I0219 09:00:35.989682 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:35.998488 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hl8ff"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.067309 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwq5\" (UniqueName: \"kubernetes.io/projected/20fa7039-8505-420a-a55f-0214b8c5fb52-kube-api-access-pcwq5\") pod \"glance-db-create-hl8ff\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.067403 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fa7039-8505-420a-a55f-0214b8c5fb52-operator-scripts\") pod \"glance-db-create-hl8ff\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.095884 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-00eb-account-create-update-h6l8c"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.096850 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.099052 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.112233 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-00eb-account-create-update-h6l8c"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.168693 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fa7039-8505-420a-a55f-0214b8c5fb52-operator-scripts\") pod \"glance-db-create-hl8ff\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.169108 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bqkm\" (UniqueName: \"kubernetes.io/projected/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-kube-api-access-6bqkm\") pod \"glance-00eb-account-create-update-h6l8c\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.169346 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwq5\" (UniqueName: \"kubernetes.io/projected/20fa7039-8505-420a-a55f-0214b8c5fb52-kube-api-access-pcwq5\") pod \"glance-db-create-hl8ff\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.169772 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-operator-scripts\") pod \"glance-00eb-account-create-update-h6l8c\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.169502 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fa7039-8505-420a-a55f-0214b8c5fb52-operator-scripts\") pod \"glance-db-create-hl8ff\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.189363 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwq5\" (UniqueName: \"kubernetes.io/projected/20fa7039-8505-420a-a55f-0214b8c5fb52-kube-api-access-pcwq5\") pod \"glance-db-create-hl8ff\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.271628 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bqkm\" (UniqueName: \"kubernetes.io/projected/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-kube-api-access-6bqkm\") pod \"glance-00eb-account-create-update-h6l8c\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.271799 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-operator-scripts\") pod \"glance-00eb-account-create-update-h6l8c\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.272478 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-operator-scripts\") pod \"glance-00eb-account-create-update-h6l8c\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.293685 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bqkm\" (UniqueName: \"kubernetes.io/projected/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-kube-api-access-6bqkm\") pod \"glance-00eb-account-create-update-h6l8c\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.348587 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.414435 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.627285 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gtrrk"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.629615 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.640884 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gtrrk"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.683031 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-operator-scripts\") pod \"keystone-db-create-gtrrk\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.683235 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6xkr\" (UniqueName: \"kubernetes.io/projected/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-kube-api-access-m6xkr\") pod \"keystone-db-create-gtrrk\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.739118 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c09a-account-create-update-dx5kw"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.741077 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.744243 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.750035 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c09a-account-create-update-dx5kw"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.785311 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jvhd\" (UniqueName: \"kubernetes.io/projected/b749ff9b-32ac-484b-a722-c42543ebc3b9-kube-api-access-8jvhd\") pod \"keystone-c09a-account-create-update-dx5kw\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.785419 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6xkr\" (UniqueName: \"kubernetes.io/projected/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-kube-api-access-m6xkr\") pod \"keystone-db-create-gtrrk\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.785476 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-operator-scripts\") pod \"keystone-db-create-gtrrk\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.785498 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b749ff9b-32ac-484b-a722-c42543ebc3b9-operator-scripts\") pod \"keystone-c09a-account-create-update-dx5kw\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.786676 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-operator-scripts\") pod \"keystone-db-create-gtrrk\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.819554 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6xkr\" (UniqueName: \"kubernetes.io/projected/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-kube-api-access-m6xkr\") pod \"keystone-db-create-gtrrk\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.842682 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.852969 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hl8ff"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.860348 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-l5kvq"] Feb 19 09:00:36 crc kubenswrapper[4675]: E0219 09:00:36.860785 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a010b2ee-7a95-45de-aa98-4898ea6cf9ec" containerName="mariadb-account-create-update" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.860804 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="a010b2ee-7a95-45de-aa98-4898ea6cf9ec" containerName="mariadb-account-create-update" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.860978 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="a010b2ee-7a95-45de-aa98-4898ea6cf9ec" containerName="mariadb-account-create-update" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.861561 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.886037 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-l5kvq"] Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.886205 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w92n\" (UniqueName: \"kubernetes.io/projected/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-kube-api-access-7w92n\") pod \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.886422 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-operator-scripts\") pod \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\" (UID: \"a010b2ee-7a95-45de-aa98-4898ea6cf9ec\") " Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.886773 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b749ff9b-32ac-484b-a722-c42543ebc3b9-operator-scripts\") pod \"keystone-c09a-account-create-update-dx5kw\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.886838 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rbnh\" (UniqueName: \"kubernetes.io/projected/1d990782-ad62-4a59-8f00-02917c2c957c-kube-api-access-7rbnh\") pod \"placement-db-create-l5kvq\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.886913 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.886957 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jvhd\" (UniqueName: \"kubernetes.io/projected/b749ff9b-32ac-484b-a722-c42543ebc3b9-kube-api-access-8jvhd\") pod \"keystone-c09a-account-create-update-dx5kw\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.887009 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d990782-ad62-4a59-8f00-02917c2c957c-operator-scripts\") pod \"placement-db-create-l5kvq\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:36 crc kubenswrapper[4675]: E0219 09:00:36.887349 4675 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 19 09:00:36 crc kubenswrapper[4675]: E0219 09:00:36.887372 4675 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 19 09:00:36 crc kubenswrapper[4675]: E0219 09:00:36.887424 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift podName:74a1ef5b-9639-489e-9dc7-2da01bffa6f2 nodeName:}" failed. No retries permitted until 2026-02-19 09:00:44.887405146 +0000 UTC m=+1026.514495414 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift") pod "swift-storage-0" (UID: "74a1ef5b-9639-489e-9dc7-2da01bffa6f2") : configmap "swift-ring-files" not found Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.887429 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a010b2ee-7a95-45de-aa98-4898ea6cf9ec" (UID: "a010b2ee-7a95-45de-aa98-4898ea6cf9ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.889556 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b749ff9b-32ac-484b-a722-c42543ebc3b9-operator-scripts\") pod \"keystone-c09a-account-create-update-dx5kw\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.890836 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-kube-api-access-7w92n" (OuterVolumeSpecName: "kube-api-access-7w92n") pod "a010b2ee-7a95-45de-aa98-4898ea6cf9ec" (UID: "a010b2ee-7a95-45de-aa98-4898ea6cf9ec"). InnerVolumeSpecName "kube-api-access-7w92n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.915414 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jvhd\" (UniqueName: \"kubernetes.io/projected/b749ff9b-32ac-484b-a722-c42543ebc3b9-kube-api-access-8jvhd\") pod \"keystone-c09a-account-create-update-dx5kw\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.956360 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.991635 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rbnh\" (UniqueName: \"kubernetes.io/projected/1d990782-ad62-4a59-8f00-02917c2c957c-kube-api-access-7rbnh\") pod \"placement-db-create-l5kvq\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.992130 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d990782-ad62-4a59-8f00-02917c2c957c-operator-scripts\") pod \"placement-db-create-l5kvq\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.992202 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.992214 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w92n\" (UniqueName: \"kubernetes.io/projected/a010b2ee-7a95-45de-aa98-4898ea6cf9ec-kube-api-access-7w92n\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:36 crc kubenswrapper[4675]: I0219 09:00:36.992948 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d990782-ad62-4a59-8f00-02917c2c957c-operator-scripts\") pod \"placement-db-create-l5kvq\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.015246 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rbnh\" (UniqueName: \"kubernetes.io/projected/1d990782-ad62-4a59-8f00-02917c2c957c-kube-api-access-7rbnh\") pod \"placement-db-create-l5kvq\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.019489 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3afe-account-create-update-ptbwn"] Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.020676 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.030953 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.031996 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3afe-account-create-update-ptbwn"] Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.074850 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.094527 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kmvz\" (UniqueName: \"kubernetes.io/projected/51845ac1-8df1-4f01-860e-c0a0d0eb0508-kube-api-access-4kmvz\") pod \"placement-3afe-account-create-update-ptbwn\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.094664 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51845ac1-8df1-4f01-860e-c0a0d0eb0508-operator-scripts\") pod \"placement-3afe-account-create-update-ptbwn\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.125506 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-00eb-account-create-update-h6l8c"] Feb 19 09:00:37 crc kubenswrapper[4675]: W0219 09:00:37.131916 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99cb40fe_72ba_4abd_bbdb_f21f7f37af3f.slice/crio-78f0f30d1dd46931fcc94950f9b7eb7f33ab6a0a66c5c2b0782b405af05d10d6 WatchSource:0}: Error finding container 78f0f30d1dd46931fcc94950f9b7eb7f33ab6a0a66c5c2b0782b405af05d10d6: Status 404 returned error can't find the container with id 78f0f30d1dd46931fcc94950f9b7eb7f33ab6a0a66c5c2b0782b405af05d10d6 Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.184398 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.197829 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51845ac1-8df1-4f01-860e-c0a0d0eb0508-operator-scripts\") pod \"placement-3afe-account-create-update-ptbwn\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.198006 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kmvz\" (UniqueName: \"kubernetes.io/projected/51845ac1-8df1-4f01-860e-c0a0d0eb0508-kube-api-access-4kmvz\") pod \"placement-3afe-account-create-update-ptbwn\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.200023 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51845ac1-8df1-4f01-860e-c0a0d0eb0508-operator-scripts\") pod \"placement-3afe-account-create-update-ptbwn\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.222931 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kmvz\" (UniqueName: \"kubernetes.io/projected/51845ac1-8df1-4f01-860e-c0a0d0eb0508-kube-api-access-4kmvz\") pod \"placement-3afe-account-create-update-ptbwn\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.391604 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.440869 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00eb-account-create-update-h6l8c" event={"ID":"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f","Type":"ContainerStarted","Data":"1a2ce06e34a659ff629585bf0c6c451d0c806958da9acd153dd3590c202efcff"} Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.440928 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00eb-account-create-update-h6l8c" event={"ID":"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f","Type":"ContainerStarted","Data":"78f0f30d1dd46931fcc94950f9b7eb7f33ab6a0a66c5c2b0782b405af05d10d6"} Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.450997 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-6lkh9" event={"ID":"a010b2ee-7a95-45de-aa98-4898ea6cf9ec","Type":"ContainerDied","Data":"95470785b4046bc386d1df97b0cf52b64a80b5e80549c17bfdc51fbc28f3b157"} Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.451054 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95470785b4046bc386d1df97b0cf52b64a80b5e80549c17bfdc51fbc28f3b157" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.451075 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6lkh9" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.465294 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-00eb-account-create-update-h6l8c" podStartSLOduration=1.465272602 podStartE2EDuration="1.465272602s" podCreationTimestamp="2026-02-19 09:00:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:00:37.460970583 +0000 UTC m=+1019.088060851" watchObservedRunningTime="2026-02-19 09:00:37.465272602 +0000 UTC m=+1019.092362870" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.467082 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hl8ff" event={"ID":"20fa7039-8505-420a-a55f-0214b8c5fb52","Type":"ContainerStarted","Data":"5a8c48c07e166c1ef8a78d91ad8dbf9f60a54c6aaa936483addda5d01b11c391"} Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.467136 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hl8ff" event={"ID":"20fa7039-8505-420a-a55f-0214b8c5fb52","Type":"ContainerStarted","Data":"6746737a1632b23ba31d27d586339fe25c95838fc6a6374f363493334ac6729a"} Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.493610 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-hl8ff" podStartSLOduration=2.493592862 podStartE2EDuration="2.493592862s" podCreationTimestamp="2026-02-19 09:00:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:00:37.48291951 +0000 UTC m=+1019.110009788" watchObservedRunningTime="2026-02-19 09:00:37.493592862 +0000 UTC m=+1019.120683130" Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.538995 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gtrrk"] Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.707412 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c09a-account-create-update-dx5kw"] Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.796954 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-l5kvq"] Feb 19 09:00:37 crc kubenswrapper[4675]: W0219 09:00:37.963021 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51845ac1_8df1_4f01_860e_c0a0d0eb0508.slice/crio-1f6c99effe7d803b3dd05d892b42ff57da3f751c94043388d2c1757a94681372 WatchSource:0}: Error finding container 1f6c99effe7d803b3dd05d892b42ff57da3f751c94043388d2c1757a94681372: Status 404 returned error can't find the container with id 1f6c99effe7d803b3dd05d892b42ff57da3f751c94043388d2c1757a94681372 Feb 19 09:00:37 crc kubenswrapper[4675]: I0219 09:00:37.965128 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3afe-account-create-update-ptbwn"] Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.253096 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.350705 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-s62wm"] Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.350986 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-s62wm" podUID="656cfb75-4638-4306-8490-1854667f58d0" containerName="dnsmasq-dns" containerID="cri-o://f6a6f3613036a1b2e8ed89e7479a9d36466e291c3e396940458f3b37fee9324e" gracePeriod=10 Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.408722 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-s62wm" podUID="656cfb75-4638-4306-8490-1854667f58d0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.483885 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3afe-account-create-update-ptbwn" event={"ID":"51845ac1-8df1-4f01-860e-c0a0d0eb0508","Type":"ContainerStarted","Data":"0e2a9adc7b5f959b07520dcde45292ed36759b337879b77300141ab7dd29ac13"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.483943 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3afe-account-create-update-ptbwn" event={"ID":"51845ac1-8df1-4f01-860e-c0a0d0eb0508","Type":"ContainerStarted","Data":"1f6c99effe7d803b3dd05d892b42ff57da3f751c94043388d2c1757a94681372"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.501446 4675 generic.go:334] "Generic (PLEG): container finished" podID="1d990782-ad62-4a59-8f00-02917c2c957c" containerID="25989343d3a8a6300cf62f853a3f65f9a5f665095adfe2e3cc15f2e25b488b8f" exitCode=0 Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.501526 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l5kvq" event={"ID":"1d990782-ad62-4a59-8f00-02917c2c957c","Type":"ContainerDied","Data":"25989343d3a8a6300cf62f853a3f65f9a5f665095adfe2e3cc15f2e25b488b8f"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.501553 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l5kvq" event={"ID":"1d990782-ad62-4a59-8f00-02917c2c957c","Type":"ContainerStarted","Data":"9591fb1b4b03f4129844b64a01b852427edcb973db7c07a35a46b360cb75a5eb"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.510924 4675 generic.go:334] "Generic (PLEG): container finished" podID="20fa7039-8505-420a-a55f-0214b8c5fb52" containerID="5a8c48c07e166c1ef8a78d91ad8dbf9f60a54c6aaa936483addda5d01b11c391" exitCode=0 Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.511021 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hl8ff" event={"ID":"20fa7039-8505-420a-a55f-0214b8c5fb52","Type":"ContainerDied","Data":"5a8c48c07e166c1ef8a78d91ad8dbf9f60a54c6aaa936483addda5d01b11c391"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.514178 4675 generic.go:334] "Generic (PLEG): container finished" podID="2de7aed7-aa20-4ba7-95dc-4d96b45c0caa" containerID="f82a995bae45cad532588f3de77fe1de7b607c5b389a6b0069a7aa4893b3a535" exitCode=0 Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.514312 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gtrrk" event={"ID":"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa","Type":"ContainerDied","Data":"f82a995bae45cad532588f3de77fe1de7b607c5b389a6b0069a7aa4893b3a535"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.514351 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gtrrk" event={"ID":"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa","Type":"ContainerStarted","Data":"f43be1c81d29cb413537ebb63f6794376aa81a58166313eb24fe4d97fd118074"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.517123 4675 generic.go:334] "Generic (PLEG): container finished" podID="656cfb75-4638-4306-8490-1854667f58d0" containerID="f6a6f3613036a1b2e8ed89e7479a9d36466e291c3e396940458f3b37fee9324e" exitCode=0 Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.517196 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-s62wm" event={"ID":"656cfb75-4638-4306-8490-1854667f58d0","Type":"ContainerDied","Data":"f6a6f3613036a1b2e8ed89e7479a9d36466e291c3e396940458f3b37fee9324e"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.526588 4675 generic.go:334] "Generic (PLEG): container finished" podID="99cb40fe-72ba-4abd-bbdb-f21f7f37af3f" containerID="1a2ce06e34a659ff629585bf0c6c451d0c806958da9acd153dd3590c202efcff" exitCode=0 Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.526666 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00eb-account-create-update-h6l8c" event={"ID":"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f","Type":"ContainerDied","Data":"1a2ce06e34a659ff629585bf0c6c451d0c806958da9acd153dd3590c202efcff"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.527996 4675 generic.go:334] "Generic (PLEG): container finished" podID="b749ff9b-32ac-484b-a722-c42543ebc3b9" containerID="b4a058cb8fa7120b3594062632f768f4f2e996b0c7cf5bcfd8129a15299b3e48" exitCode=0 Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.528033 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c09a-account-create-update-dx5kw" event={"ID":"b749ff9b-32ac-484b-a722-c42543ebc3b9","Type":"ContainerDied","Data":"b4a058cb8fa7120b3594062632f768f4f2e996b0c7cf5bcfd8129a15299b3e48"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.528057 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c09a-account-create-update-dx5kw" event={"ID":"b749ff9b-32ac-484b-a722-c42543ebc3b9","Type":"ContainerStarted","Data":"0395b4d46e24a77c7dfcc8605c909823d44fe531b3bc18dd3771927f1a331f93"} Feb 19 09:00:38 crc kubenswrapper[4675]: I0219 09:00:38.982253 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.051339 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2c5b\" (UniqueName: \"kubernetes.io/projected/656cfb75-4638-4306-8490-1854667f58d0-kube-api-access-t2c5b\") pod \"656cfb75-4638-4306-8490-1854667f58d0\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.051408 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-nb\") pod \"656cfb75-4638-4306-8490-1854667f58d0\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.051540 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-dns-svc\") pod \"656cfb75-4638-4306-8490-1854667f58d0\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.051571 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-sb\") pod \"656cfb75-4638-4306-8490-1854667f58d0\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.051612 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-config\") pod \"656cfb75-4638-4306-8490-1854667f58d0\" (UID: \"656cfb75-4638-4306-8490-1854667f58d0\") " Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.073002 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656cfb75-4638-4306-8490-1854667f58d0-kube-api-access-t2c5b" (OuterVolumeSpecName: "kube-api-access-t2c5b") pod "656cfb75-4638-4306-8490-1854667f58d0" (UID: "656cfb75-4638-4306-8490-1854667f58d0"). InnerVolumeSpecName "kube-api-access-t2c5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.148971 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-6lkh9"] Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.149007 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-6lkh9"] Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.154928 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2c5b\" (UniqueName: \"kubernetes.io/projected/656cfb75-4638-4306-8490-1854667f58d0-kube-api-access-t2c5b\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.159767 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "656cfb75-4638-4306-8490-1854667f58d0" (UID: "656cfb75-4638-4306-8490-1854667f58d0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.169168 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-config" (OuterVolumeSpecName: "config") pod "656cfb75-4638-4306-8490-1854667f58d0" (UID: "656cfb75-4638-4306-8490-1854667f58d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.176110 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "656cfb75-4638-4306-8490-1854667f58d0" (UID: "656cfb75-4638-4306-8490-1854667f58d0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.196811 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "656cfb75-4638-4306-8490-1854667f58d0" (UID: "656cfb75-4638-4306-8490-1854667f58d0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.256570 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.256622 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.256652 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.256663 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/656cfb75-4638-4306-8490-1854667f58d0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.548923 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-s62wm" event={"ID":"656cfb75-4638-4306-8490-1854667f58d0","Type":"ContainerDied","Data":"5b838a9fb800e92c497801db6d9a7fc09cbc67ffc614d5fdf7a3be2031ab8b22"} Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.548978 4675 scope.go:117] "RemoveContainer" containerID="f6a6f3613036a1b2e8ed89e7479a9d36466e291c3e396940458f3b37fee9324e" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.549140 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-s62wm" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.554005 4675 generic.go:334] "Generic (PLEG): container finished" podID="51845ac1-8df1-4f01-860e-c0a0d0eb0508" containerID="0e2a9adc7b5f959b07520dcde45292ed36759b337879b77300141ab7dd29ac13" exitCode=0 Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.554331 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3afe-account-create-update-ptbwn" event={"ID":"51845ac1-8df1-4f01-860e-c0a0d0eb0508","Type":"ContainerDied","Data":"0e2a9adc7b5f959b07520dcde45292ed36759b337879b77300141ab7dd29ac13"} Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.623047 4675 scope.go:117] "RemoveContainer" containerID="9f6e34c1494f033243540ebd9f9c3d53814851da486fd1c429ea5292b803b6c1" Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.635969 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-s62wm"] Feb 19 09:00:39 crc kubenswrapper[4675]: I0219 09:00:39.648340 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-s62wm"] Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.078036 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.176951 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fa7039-8505-420a-a55f-0214b8c5fb52-operator-scripts\") pod \"20fa7039-8505-420a-a55f-0214b8c5fb52\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.177133 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcwq5\" (UniqueName: \"kubernetes.io/projected/20fa7039-8505-420a-a55f-0214b8c5fb52-kube-api-access-pcwq5\") pod \"20fa7039-8505-420a-a55f-0214b8c5fb52\" (UID: \"20fa7039-8505-420a-a55f-0214b8c5fb52\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.179868 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20fa7039-8505-420a-a55f-0214b8c5fb52-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20fa7039-8505-420a-a55f-0214b8c5fb52" (UID: "20fa7039-8505-420a-a55f-0214b8c5fb52"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.200971 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20fa7039-8505-420a-a55f-0214b8c5fb52-kube-api-access-pcwq5" (OuterVolumeSpecName: "kube-api-access-pcwq5") pod "20fa7039-8505-420a-a55f-0214b8c5fb52" (UID: "20fa7039-8505-420a-a55f-0214b8c5fb52"). InnerVolumeSpecName "kube-api-access-pcwq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.285934 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20fa7039-8505-420a-a55f-0214b8c5fb52-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.285987 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcwq5\" (UniqueName: \"kubernetes.io/projected/20fa7039-8505-420a-a55f-0214b8c5fb52-kube-api-access-pcwq5\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.369251 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.377563 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.387040 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.406141 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.412260 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488019 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kmvz\" (UniqueName: \"kubernetes.io/projected/51845ac1-8df1-4f01-860e-c0a0d0eb0508-kube-api-access-4kmvz\") pod \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488123 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jvhd\" (UniqueName: \"kubernetes.io/projected/b749ff9b-32ac-484b-a722-c42543ebc3b9-kube-api-access-8jvhd\") pod \"b749ff9b-32ac-484b-a722-c42543ebc3b9\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488201 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b749ff9b-32ac-484b-a722-c42543ebc3b9-operator-scripts\") pod \"b749ff9b-32ac-484b-a722-c42543ebc3b9\" (UID: \"b749ff9b-32ac-484b-a722-c42543ebc3b9\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488239 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bqkm\" (UniqueName: \"kubernetes.io/projected/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-kube-api-access-6bqkm\") pod \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488335 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-operator-scripts\") pod \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\" (UID: \"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488360 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6xkr\" (UniqueName: \"kubernetes.io/projected/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-kube-api-access-m6xkr\") pod \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488383 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51845ac1-8df1-4f01-860e-c0a0d0eb0508-operator-scripts\") pod \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\" (UID: \"51845ac1-8df1-4f01-860e-c0a0d0eb0508\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488444 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-operator-scripts\") pod \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\" (UID: \"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488476 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rbnh\" (UniqueName: \"kubernetes.io/projected/1d990782-ad62-4a59-8f00-02917c2c957c-kube-api-access-7rbnh\") pod \"1d990782-ad62-4a59-8f00-02917c2c957c\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.488511 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d990782-ad62-4a59-8f00-02917c2c957c-operator-scripts\") pod \"1d990782-ad62-4a59-8f00-02917c2c957c\" (UID: \"1d990782-ad62-4a59-8f00-02917c2c957c\") " Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.489943 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d990782-ad62-4a59-8f00-02917c2c957c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d990782-ad62-4a59-8f00-02917c2c957c" (UID: "1d990782-ad62-4a59-8f00-02917c2c957c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.490843 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99cb40fe-72ba-4abd-bbdb-f21f7f37af3f" (UID: "99cb40fe-72ba-4abd-bbdb-f21f7f37af3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.491018 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51845ac1-8df1-4f01-860e-c0a0d0eb0508-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51845ac1-8df1-4f01-860e-c0a0d0eb0508" (UID: "51845ac1-8df1-4f01-860e-c0a0d0eb0508"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.494819 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b749ff9b-32ac-484b-a722-c42543ebc3b9-kube-api-access-8jvhd" (OuterVolumeSpecName: "kube-api-access-8jvhd") pod "b749ff9b-32ac-484b-a722-c42543ebc3b9" (UID: "b749ff9b-32ac-484b-a722-c42543ebc3b9"). InnerVolumeSpecName "kube-api-access-8jvhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.495183 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b749ff9b-32ac-484b-a722-c42543ebc3b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b749ff9b-32ac-484b-a722-c42543ebc3b9" (UID: "b749ff9b-32ac-484b-a722-c42543ebc3b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.495697 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2de7aed7-aa20-4ba7-95dc-4d96b45c0caa" (UID: "2de7aed7-aa20-4ba7-95dc-4d96b45c0caa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.528887 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d990782-ad62-4a59-8f00-02917c2c957c-kube-api-access-7rbnh" (OuterVolumeSpecName: "kube-api-access-7rbnh") pod "1d990782-ad62-4a59-8f00-02917c2c957c" (UID: "1d990782-ad62-4a59-8f00-02917c2c957c"). InnerVolumeSpecName "kube-api-access-7rbnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.528973 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51845ac1-8df1-4f01-860e-c0a0d0eb0508-kube-api-access-4kmvz" (OuterVolumeSpecName: "kube-api-access-4kmvz") pod "51845ac1-8df1-4f01-860e-c0a0d0eb0508" (UID: "51845ac1-8df1-4f01-860e-c0a0d0eb0508"). InnerVolumeSpecName "kube-api-access-4kmvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.531326 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-kube-api-access-m6xkr" (OuterVolumeSpecName: "kube-api-access-m6xkr") pod "2de7aed7-aa20-4ba7-95dc-4d96b45c0caa" (UID: "2de7aed7-aa20-4ba7-95dc-4d96b45c0caa"). InnerVolumeSpecName "kube-api-access-m6xkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.532194 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-kube-api-access-6bqkm" (OuterVolumeSpecName: "kube-api-access-6bqkm") pod "99cb40fe-72ba-4abd-bbdb-f21f7f37af3f" (UID: "99cb40fe-72ba-4abd-bbdb-f21f7f37af3f"). InnerVolumeSpecName "kube-api-access-6bqkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.581845 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c09a-account-create-update-dx5kw" event={"ID":"b749ff9b-32ac-484b-a722-c42543ebc3b9","Type":"ContainerDied","Data":"0395b4d46e24a77c7dfcc8605c909823d44fe531b3bc18dd3771927f1a331f93"} Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.581898 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0395b4d46e24a77c7dfcc8605c909823d44fe531b3bc18dd3771927f1a331f93" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.581969 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c09a-account-create-update-dx5kw" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591527 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591595 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rbnh\" (UniqueName: \"kubernetes.io/projected/1d990782-ad62-4a59-8f00-02917c2c957c-kube-api-access-7rbnh\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591617 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d990782-ad62-4a59-8f00-02917c2c957c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591627 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kmvz\" (UniqueName: \"kubernetes.io/projected/51845ac1-8df1-4f01-860e-c0a0d0eb0508-kube-api-access-4kmvz\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591660 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jvhd\" (UniqueName: \"kubernetes.io/projected/b749ff9b-32ac-484b-a722-c42543ebc3b9-kube-api-access-8jvhd\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591671 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b749ff9b-32ac-484b-a722-c42543ebc3b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591680 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bqkm\" (UniqueName: \"kubernetes.io/projected/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-kube-api-access-6bqkm\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591689 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591698 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6xkr\" (UniqueName: \"kubernetes.io/projected/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa-kube-api-access-m6xkr\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.591707 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51845ac1-8df1-4f01-860e-c0a0d0eb0508-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.601243 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3afe-account-create-update-ptbwn" event={"ID":"51845ac1-8df1-4f01-860e-c0a0d0eb0508","Type":"ContainerDied","Data":"1f6c99effe7d803b3dd05d892b42ff57da3f751c94043388d2c1757a94681372"} Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.601321 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f6c99effe7d803b3dd05d892b42ff57da3f751c94043388d2c1757a94681372" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.601395 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3afe-account-create-update-ptbwn" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.617041 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gtrrk" event={"ID":"2de7aed7-aa20-4ba7-95dc-4d96b45c0caa","Type":"ContainerDied","Data":"f43be1c81d29cb413537ebb63f6794376aa81a58166313eb24fe4d97fd118074"} Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.617397 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f43be1c81d29cb413537ebb63f6794376aa81a58166313eb24fe4d97fd118074" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.617495 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gtrrk" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.626065 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-00eb-account-create-update-h6l8c" event={"ID":"99cb40fe-72ba-4abd-bbdb-f21f7f37af3f","Type":"ContainerDied","Data":"78f0f30d1dd46931fcc94950f9b7eb7f33ab6a0a66c5c2b0782b405af05d10d6"} Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.626108 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f0f30d1dd46931fcc94950f9b7eb7f33ab6a0a66c5c2b0782b405af05d10d6" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.626174 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-00eb-account-create-update-h6l8c" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.641221 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l5kvq" event={"ID":"1d990782-ad62-4a59-8f00-02917c2c957c","Type":"ContainerDied","Data":"9591fb1b4b03f4129844b64a01b852427edcb973db7c07a35a46b360cb75a5eb"} Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.641577 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9591fb1b4b03f4129844b64a01b852427edcb973db7c07a35a46b360cb75a5eb" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.641797 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l5kvq" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.652916 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hl8ff" event={"ID":"20fa7039-8505-420a-a55f-0214b8c5fb52","Type":"ContainerDied","Data":"6746737a1632b23ba31d27d586339fe25c95838fc6a6374f363493334ac6729a"} Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.652965 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6746737a1632b23ba31d27d586339fe25c95838fc6a6374f363493334ac6729a" Feb 19 09:00:40 crc kubenswrapper[4675]: I0219 09:00:40.653039 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hl8ff" Feb 19 09:00:41 crc kubenswrapper[4675]: I0219 09:00:41.115868 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="656cfb75-4638-4306-8490-1854667f58d0" path="/var/lib/kubelet/pods/656cfb75-4638-4306-8490-1854667f58d0/volumes" Feb 19 09:00:41 crc kubenswrapper[4675]: I0219 09:00:41.117195 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a010b2ee-7a95-45de-aa98-4898ea6cf9ec" path="/var/lib/kubelet/pods/a010b2ee-7a95-45de-aa98-4898ea6cf9ec/volumes" Feb 19 09:00:41 crc kubenswrapper[4675]: I0219 09:00:41.743623 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:00:41 crc kubenswrapper[4675]: I0219 09:00:41.744201 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.873773 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-q8w87"] Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874105 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d990782-ad62-4a59-8f00-02917c2c957c" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874117 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d990782-ad62-4a59-8f00-02917c2c957c" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874164 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656cfb75-4638-4306-8490-1854667f58d0" containerName="dnsmasq-dns" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874172 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="656cfb75-4638-4306-8490-1854667f58d0" containerName="dnsmasq-dns" Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874186 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20fa7039-8505-420a-a55f-0214b8c5fb52" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874194 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="20fa7039-8505-420a-a55f-0214b8c5fb52" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874204 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656cfb75-4638-4306-8490-1854667f58d0" containerName="init" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874210 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="656cfb75-4638-4306-8490-1854667f58d0" containerName="init" Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874220 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99cb40fe-72ba-4abd-bbdb-f21f7f37af3f" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874227 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="99cb40fe-72ba-4abd-bbdb-f21f7f37af3f" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874237 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51845ac1-8df1-4f01-860e-c0a0d0eb0508" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874243 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="51845ac1-8df1-4f01-860e-c0a0d0eb0508" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874257 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de7aed7-aa20-4ba7-95dc-4d96b45c0caa" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874264 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de7aed7-aa20-4ba7-95dc-4d96b45c0caa" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: E0219 09:00:42.874276 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b749ff9b-32ac-484b-a722-c42543ebc3b9" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874682 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b749ff9b-32ac-484b-a722-c42543ebc3b9" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874836 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="b749ff9b-32ac-484b-a722-c42543ebc3b9" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874844 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d990782-ad62-4a59-8f00-02917c2c957c" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874861 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="20fa7039-8505-420a-a55f-0214b8c5fb52" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874871 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="656cfb75-4638-4306-8490-1854667f58d0" containerName="dnsmasq-dns" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874878 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="51845ac1-8df1-4f01-860e-c0a0d0eb0508" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874886 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de7aed7-aa20-4ba7-95dc-4d96b45c0caa" containerName="mariadb-database-create" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.874896 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="99cb40fe-72ba-4abd-bbdb-f21f7f37af3f" containerName="mariadb-account-create-update" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.875382 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.878374 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.892061 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-q8w87"] Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.929582 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6999\" (UniqueName: \"kubernetes.io/projected/3da34c71-e247-423f-be5a-5307a7eaab65-kube-api-access-c6999\") pod \"root-account-create-update-q8w87\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:42 crc kubenswrapper[4675]: I0219 09:00:42.929699 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da34c71-e247-423f-be5a-5307a7eaab65-operator-scripts\") pod \"root-account-create-update-q8w87\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:43 crc kubenswrapper[4675]: I0219 09:00:43.031236 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da34c71-e247-423f-be5a-5307a7eaab65-operator-scripts\") pod \"root-account-create-update-q8w87\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:43 crc kubenswrapper[4675]: I0219 09:00:43.031404 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6999\" (UniqueName: \"kubernetes.io/projected/3da34c71-e247-423f-be5a-5307a7eaab65-kube-api-access-c6999\") pod \"root-account-create-update-q8w87\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:43 crc kubenswrapper[4675]: I0219 09:00:43.032241 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da34c71-e247-423f-be5a-5307a7eaab65-operator-scripts\") pod \"root-account-create-update-q8w87\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:43 crc kubenswrapper[4675]: I0219 09:00:43.049593 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6999\" (UniqueName: \"kubernetes.io/projected/3da34c71-e247-423f-be5a-5307a7eaab65-kube-api-access-c6999\") pod \"root-account-create-update-q8w87\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:43 crc kubenswrapper[4675]: I0219 09:00:43.252724 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:43 crc kubenswrapper[4675]: I0219 09:00:43.739355 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-q8w87"] Feb 19 09:00:43 crc kubenswrapper[4675]: W0219 09:00:43.741544 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3da34c71_e247_423f_be5a_5307a7eaab65.slice/crio-e21452e49fbae1910c2944ec45cd75a6a2ba91f7ce045a70aedb50ce3a04e3ef WatchSource:0}: Error finding container e21452e49fbae1910c2944ec45cd75a6a2ba91f7ce045a70aedb50ce3a04e3ef: Status 404 returned error can't find the container with id e21452e49fbae1910c2944ec45cd75a6a2ba91f7ce045a70aedb50ce3a04e3ef Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.065571 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.680613 4675 generic.go:334] "Generic (PLEG): container finished" podID="dadf56d2-1dd1-4590-8fd7-0cc266af5377" containerID="f3b6b52a893e965134b4ccedd11807dfd49efb1f4e73f10e2cc0ec21ae2c2b40" exitCode=0 Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.680674 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s4x6m" event={"ID":"dadf56d2-1dd1-4590-8fd7-0cc266af5377","Type":"ContainerDied","Data":"f3b6b52a893e965134b4ccedd11807dfd49efb1f4e73f10e2cc0ec21ae2c2b40"} Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.682482 4675 generic.go:334] "Generic (PLEG): container finished" podID="3da34c71-e247-423f-be5a-5307a7eaab65" containerID="27e6b292808a508110eee774e6ba38c1d382dac63eda9b8635c1c8d7aa0f15b5" exitCode=0 Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.682543 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-q8w87" event={"ID":"3da34c71-e247-423f-be5a-5307a7eaab65","Type":"ContainerDied","Data":"27e6b292808a508110eee774e6ba38c1d382dac63eda9b8635c1c8d7aa0f15b5"} Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.682732 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-q8w87" event={"ID":"3da34c71-e247-423f-be5a-5307a7eaab65","Type":"ContainerStarted","Data":"e21452e49fbae1910c2944ec45cd75a6a2ba91f7ce045a70aedb50ce3a04e3ef"} Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.967087 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.975230 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74a1ef5b-9639-489e-9dc7-2da01bffa6f2-etc-swift\") pod \"swift-storage-0\" (UID: \"74a1ef5b-9639-489e-9dc7-2da01bffa6f2\") " pod="openstack/swift-storage-0" Feb 19 09:00:44 crc kubenswrapper[4675]: I0219 09:00:44.977783 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 19 09:00:45 crc kubenswrapper[4675]: I0219 09:00:45.567824 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 19 09:00:45 crc kubenswrapper[4675]: I0219 09:00:45.689015 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"9974c26dfafae66ab69c43697037fd0a2655c7f9b6925a6f0c0c4bebb5a5b35c"} Feb 19 09:00:45 crc kubenswrapper[4675]: I0219 09:00:45.690076 4675 generic.go:334] "Generic (PLEG): container finished" podID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerID="a993f5ff21a91f8f3b65ffb590d4332551df482a908b1a402992186078332313" exitCode=0 Feb 19 09:00:45 crc kubenswrapper[4675]: I0219 09:00:45.690146 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1adde38f-79d6-4ee4-82ed-97004ffb7f85","Type":"ContainerDied","Data":"a993f5ff21a91f8f3b65ffb590d4332551df482a908b1a402992186078332313"} Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.139631 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.147060 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.198582 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dadf56d2-1dd1-4590-8fd7-0cc266af5377-etc-swift\") pod \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.198661 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-combined-ca-bundle\") pod \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.198738 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da34c71-e247-423f-be5a-5307a7eaab65-operator-scripts\") pod \"3da34c71-e247-423f-be5a-5307a7eaab65\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.198777 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jqgc\" (UniqueName: \"kubernetes.io/projected/dadf56d2-1dd1-4590-8fd7-0cc266af5377-kube-api-access-6jqgc\") pod \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.198798 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6999\" (UniqueName: \"kubernetes.io/projected/3da34c71-e247-423f-be5a-5307a7eaab65-kube-api-access-c6999\") pod \"3da34c71-e247-423f-be5a-5307a7eaab65\" (UID: \"3da34c71-e247-423f-be5a-5307a7eaab65\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.199393 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-ring-data-devices\") pod \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.199173 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3da34c71-e247-423f-be5a-5307a7eaab65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3da34c71-e247-423f-be5a-5307a7eaab65" (UID: "3da34c71-e247-423f-be5a-5307a7eaab65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.199456 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-scripts\") pod \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.199485 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-swiftconf\") pod \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.199543 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-dispersionconf\") pod \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\" (UID: \"dadf56d2-1dd1-4590-8fd7-0cc266af5377\") " Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.199689 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dadf56d2-1dd1-4590-8fd7-0cc266af5377-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "dadf56d2-1dd1-4590-8fd7-0cc266af5377" (UID: "dadf56d2-1dd1-4590-8fd7-0cc266af5377"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.199979 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "dadf56d2-1dd1-4590-8fd7-0cc266af5377" (UID: "dadf56d2-1dd1-4590-8fd7-0cc266af5377"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.200133 4675 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dadf56d2-1dd1-4590-8fd7-0cc266af5377-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.200154 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da34c71-e247-423f-be5a-5307a7eaab65-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.200168 4675 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.204881 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dadf56d2-1dd1-4590-8fd7-0cc266af5377-kube-api-access-6jqgc" (OuterVolumeSpecName: "kube-api-access-6jqgc") pod "dadf56d2-1dd1-4590-8fd7-0cc266af5377" (UID: "dadf56d2-1dd1-4590-8fd7-0cc266af5377"). InnerVolumeSpecName "kube-api-access-6jqgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.206332 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da34c71-e247-423f-be5a-5307a7eaab65-kube-api-access-c6999" (OuterVolumeSpecName: "kube-api-access-c6999") pod "3da34c71-e247-423f-be5a-5307a7eaab65" (UID: "3da34c71-e247-423f-be5a-5307a7eaab65"). InnerVolumeSpecName "kube-api-access-c6999". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.210270 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "dadf56d2-1dd1-4590-8fd7-0cc266af5377" (UID: "dadf56d2-1dd1-4590-8fd7-0cc266af5377"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.237712 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-scripts" (OuterVolumeSpecName: "scripts") pod "dadf56d2-1dd1-4590-8fd7-0cc266af5377" (UID: "dadf56d2-1dd1-4590-8fd7-0cc266af5377"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.246800 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "dadf56d2-1dd1-4590-8fd7-0cc266af5377" (UID: "dadf56d2-1dd1-4590-8fd7-0cc266af5377"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.267823 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dadf56d2-1dd1-4590-8fd7-0cc266af5377" (UID: "dadf56d2-1dd1-4590-8fd7-0cc266af5377"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.276132 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-g8ngb"] Feb 19 09:00:46 crc kubenswrapper[4675]: E0219 09:00:46.276586 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dadf56d2-1dd1-4590-8fd7-0cc266af5377" containerName="swift-ring-rebalance" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.276614 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="dadf56d2-1dd1-4590-8fd7-0cc266af5377" containerName="swift-ring-rebalance" Feb 19 09:00:46 crc kubenswrapper[4675]: E0219 09:00:46.276652 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da34c71-e247-423f-be5a-5307a7eaab65" containerName="mariadb-account-create-update" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.276663 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da34c71-e247-423f-be5a-5307a7eaab65" containerName="mariadb-account-create-update" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.276865 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="dadf56d2-1dd1-4590-8fd7-0cc266af5377" containerName="swift-ring-rebalance" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.276903 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="3da34c71-e247-423f-be5a-5307a7eaab65" containerName="mariadb-account-create-update" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.277669 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.284894 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mswnr" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.286355 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.292055 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g8ngb"] Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302574 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-db-sync-config-data\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302681 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxm6b\" (UniqueName: \"kubernetes.io/projected/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-kube-api-access-cxm6b\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302707 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-config-data\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302750 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-combined-ca-bundle\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302794 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302806 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jqgc\" (UniqueName: \"kubernetes.io/projected/dadf56d2-1dd1-4590-8fd7-0cc266af5377-kube-api-access-6jqgc\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302818 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6999\" (UniqueName: \"kubernetes.io/projected/3da34c71-e247-423f-be5a-5307a7eaab65-kube-api-access-c6999\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302827 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dadf56d2-1dd1-4590-8fd7-0cc266af5377-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302835 4675 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.302843 4675 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dadf56d2-1dd1-4590-8fd7-0cc266af5377-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.404001 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxm6b\" (UniqueName: \"kubernetes.io/projected/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-kube-api-access-cxm6b\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.404046 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-config-data\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.404086 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-combined-ca-bundle\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.404141 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-db-sync-config-data\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.407503 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-config-data\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.408936 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-db-sync-config-data\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.409075 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-combined-ca-bundle\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.420480 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxm6b\" (UniqueName: \"kubernetes.io/projected/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-kube-api-access-cxm6b\") pod \"glance-db-sync-g8ngb\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.609921 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g8ngb" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.701038 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-s4x6m" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.701030 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-s4x6m" event={"ID":"dadf56d2-1dd1-4590-8fd7-0cc266af5377","Type":"ContainerDied","Data":"d358b5269eb79e735682987356149ac77d413989bdfde11915a906153765bb7d"} Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.701551 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d358b5269eb79e735682987356149ac77d413989bdfde11915a906153765bb7d" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.702494 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-q8w87" event={"ID":"3da34c71-e247-423f-be5a-5307a7eaab65","Type":"ContainerDied","Data":"e21452e49fbae1910c2944ec45cd75a6a2ba91f7ce045a70aedb50ce3a04e3ef"} Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.702556 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e21452e49fbae1910c2944ec45cd75a6a2ba91f7ce045a70aedb50ce3a04e3ef" Feb 19 09:00:46 crc kubenswrapper[4675]: I0219 09:00:46.702686 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-q8w87" Feb 19 09:00:47 crc kubenswrapper[4675]: I0219 09:00:47.230592 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g8ngb"] Feb 19 09:00:47 crc kubenswrapper[4675]: W0219 09:00:47.638049 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ee47cd5_e516_4da9_ac5f_40cbedfe103b.slice/crio-fdb9173f64fca44379a955831c660b3c255a53fdb9cb2398723e5b327f952032 WatchSource:0}: Error finding container fdb9173f64fca44379a955831c660b3c255a53fdb9cb2398723e5b327f952032: Status 404 returned error can't find the container with id fdb9173f64fca44379a955831c660b3c255a53fdb9cb2398723e5b327f952032 Feb 19 09:00:47 crc kubenswrapper[4675]: I0219 09:00:47.713317 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1adde38f-79d6-4ee4-82ed-97004ffb7f85","Type":"ContainerStarted","Data":"db61585170f749410d45a8cfd2979a86557f0fc24ab094b737b4fa31e9bf3613"} Feb 19 09:00:47 crc kubenswrapper[4675]: I0219 09:00:47.713694 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:00:47 crc kubenswrapper[4675]: I0219 09:00:47.718427 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g8ngb" event={"ID":"3ee47cd5-e516-4da9-ac5f-40cbedfe103b","Type":"ContainerStarted","Data":"fdb9173f64fca44379a955831c660b3c255a53fdb9cb2398723e5b327f952032"} Feb 19 09:00:47 crc kubenswrapper[4675]: I0219 09:00:47.750261 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=44.83791697 podStartE2EDuration="1m6.750237934s" podCreationTimestamp="2026-02-19 08:59:41 +0000 UTC" firstStartedPulling="2026-02-19 08:59:43.652715325 +0000 UTC m=+965.279805593" lastFinishedPulling="2026-02-19 09:00:05.565036289 +0000 UTC m=+987.192126557" observedRunningTime="2026-02-19 09:00:47.74612691 +0000 UTC m=+1029.373217178" watchObservedRunningTime="2026-02-19 09:00:47.750237934 +0000 UTC m=+1029.377328202" Feb 19 09:00:48 crc kubenswrapper[4675]: I0219 09:00:48.743140 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"bbb6d28c8e6b222669c14e7e1adb8ce1fb0731a0c051213bf3d485f861941c69"} Feb 19 09:00:48 crc kubenswrapper[4675]: I0219 09:00:48.743612 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"d058ad03d0cd4e05b4b8f2823928c56c67eb5485aa1cd319dbe5508e4a281f90"} Feb 19 09:00:48 crc kubenswrapper[4675]: I0219 09:00:48.743628 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"af77cc47d88a4918ab9d79690125f3c901a4f18db651a0270305a1bfea03386d"} Feb 19 09:00:48 crc kubenswrapper[4675]: I0219 09:00:48.743654 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"80f8c4256be349efd3df3a565d79bf1e4ecc2b369237b531361cd5a059ff9dce"} Feb 19 09:00:49 crc kubenswrapper[4675]: I0219 09:00:49.122328 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-q8w87"] Feb 19 09:00:49 crc kubenswrapper[4675]: I0219 09:00:49.129595 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-q8w87"] Feb 19 09:00:50 crc kubenswrapper[4675]: I0219 09:00:50.767572 4675 generic.go:334] "Generic (PLEG): container finished" podID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerID="897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b" exitCode=0 Feb 19 09:00:50 crc kubenswrapper[4675]: I0219 09:00:50.767702 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93dbb123-819d-44f7-80f5-8c38c639f5f4","Type":"ContainerDied","Data":"897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b"} Feb 19 09:00:50 crc kubenswrapper[4675]: I0219 09:00:50.774343 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"2a28ff2764dbda2d07c7687eb4b817a12d00a9735b635c4841ae830eb41a3f6d"} Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.084220 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vlnhc" podUID="0b11fc83-5e56-43d3-a340-80686a5bdf35" containerName="ovn-controller" probeResult="failure" output=< Feb 19 09:00:51 crc kubenswrapper[4675]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 19 09:00:51 crc kubenswrapper[4675]: > Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.113278 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3da34c71-e247-423f-be5a-5307a7eaab65" path="/var/lib/kubelet/pods/3da34c71-e247-423f-be5a-5307a7eaab65/volumes" Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.355546 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-s5dht" Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.409699 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-s5dht" Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.785431 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93dbb123-819d-44f7-80f5-8c38c639f5f4","Type":"ContainerStarted","Data":"f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e"} Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.787005 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.795367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"0fc292fd679937be68a23d62e19a03a037cf54e95145bda51252e50a2b64207a"} Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.795465 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"3fccc8bf5a270b6ebcd0e702798e91e356620aa7b444891ee777009e26416f63"} Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.879732 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371965.975073 podStartE2EDuration="1m10.87970341s" podCreationTimestamp="2026-02-19 08:59:41 +0000 UTC" firstStartedPulling="2026-02-19 08:59:43.404533025 +0000 UTC m=+965.031623293" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:00:51.822179008 +0000 UTC m=+1033.449269276" watchObservedRunningTime="2026-02-19 09:00:51.87970341 +0000 UTC m=+1033.506793678" Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.881581 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vlnhc-config-5cj4n"] Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.883138 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.884546 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 19 09:00:51 crc kubenswrapper[4675]: I0219 09:00:51.895367 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vlnhc-config-5cj4n"] Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.015903 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-scripts\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.015972 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-log-ovn\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.016039 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run-ovn\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.016088 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-additional-scripts\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.016110 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.016152 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69n4j\" (UniqueName: \"kubernetes.io/projected/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-kube-api-access-69n4j\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118149 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-log-ovn\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118251 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run-ovn\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118302 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-additional-scripts\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118327 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118371 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69n4j\" (UniqueName: \"kubernetes.io/projected/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-kube-api-access-69n4j\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118440 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-scripts\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118539 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118559 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run-ovn\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.118850 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-log-ovn\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.119211 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-additional-scripts\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.120772 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-scripts\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.145594 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69n4j\" (UniqueName: \"kubernetes.io/projected/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-kube-api-access-69n4j\") pod \"ovn-controller-vlnhc-config-5cj4n\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.250087 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.823880 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"0acfbeb022481fff6d5a07044d62fa0788cd99e4c360c3acf4c651aaa6ac33f8"} Feb 19 09:00:52 crc kubenswrapper[4675]: I0219 09:00:52.888282 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vlnhc-config-5cj4n"] Feb 19 09:00:52 crc kubenswrapper[4675]: W0219 09:00:52.901311 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1727efd3_5377_4922_a7cb_4b3e9b0eabdb.slice/crio-b69a030444321404b42076478306cf0a37382c8ee90622a8f43e5ed923e94db3 WatchSource:0}: Error finding container b69a030444321404b42076478306cf0a37382c8ee90622a8f43e5ed923e94db3: Status 404 returned error can't find the container with id b69a030444321404b42076478306cf0a37382c8ee90622a8f43e5ed923e94db3 Feb 19 09:00:53 crc kubenswrapper[4675]: I0219 09:00:53.833366 4675 generic.go:334] "Generic (PLEG): container finished" podID="1727efd3-5377-4922-a7cb-4b3e9b0eabdb" containerID="b801f0662ccb083ec945731718f55feff253d67d407722cdf02706cd5333269d" exitCode=0 Feb 19 09:00:53 crc kubenswrapper[4675]: I0219 09:00:53.833448 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc-config-5cj4n" event={"ID":"1727efd3-5377-4922-a7cb-4b3e9b0eabdb","Type":"ContainerDied","Data":"b801f0662ccb083ec945731718f55feff253d67d407722cdf02706cd5333269d"} Feb 19 09:00:53 crc kubenswrapper[4675]: I0219 09:00:53.833882 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc-config-5cj4n" event={"ID":"1727efd3-5377-4922-a7cb-4b3e9b0eabdb","Type":"ContainerStarted","Data":"b69a030444321404b42076478306cf0a37382c8ee90622a8f43e5ed923e94db3"} Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.137032 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-4bphc"] Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.137970 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.142090 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.160021 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4bphc"] Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.251793 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgjs5\" (UniqueName: \"kubernetes.io/projected/fd86c751-bf59-47e4-81c4-dfaeb157fa76-kube-api-access-tgjs5\") pod \"root-account-create-update-4bphc\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.251911 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd86c751-bf59-47e4-81c4-dfaeb157fa76-operator-scripts\") pod \"root-account-create-update-4bphc\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.353457 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgjs5\" (UniqueName: \"kubernetes.io/projected/fd86c751-bf59-47e4-81c4-dfaeb157fa76-kube-api-access-tgjs5\") pod \"root-account-create-update-4bphc\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.353549 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd86c751-bf59-47e4-81c4-dfaeb157fa76-operator-scripts\") pod \"root-account-create-update-4bphc\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.354762 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd86c751-bf59-47e4-81c4-dfaeb157fa76-operator-scripts\") pod \"root-account-create-update-4bphc\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.381324 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgjs5\" (UniqueName: \"kubernetes.io/projected/fd86c751-bf59-47e4-81c4-dfaeb157fa76-kube-api-access-tgjs5\") pod \"root-account-create-update-4bphc\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.562223 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4bphc" Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.863082 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"4798ab38e4f7d3fc637af754de579843557dff70a629c572f72427b482afbb12"} Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.863422 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"edde9d0101b554b2e6ed06bf2e06567efaf401de178d7643fe0dc4b80fe916a5"} Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.863433 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"f0c4970cc5ec9c113a873908bc99cd413d283924ac669b7782d4667efc221f7f"} Feb 19 09:00:54 crc kubenswrapper[4675]: I0219 09:00:54.863444 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"0c526300623e44dd3a4f274995603aebcf88568532fce0833e3f7a3e3b56c2e2"} Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.323890 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4bphc"] Feb 19 09:00:55 crc kubenswrapper[4675]: W0219 09:00:55.355722 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd86c751_bf59_47e4_81c4_dfaeb157fa76.slice/crio-9e7d59b105b9c40a9518bbe958058df52423b595d7d9a293b71f8fe214f7bf88 WatchSource:0}: Error finding container 9e7d59b105b9c40a9518bbe958058df52423b595d7d9a293b71f8fe214f7bf88: Status 404 returned error can't find the container with id 9e7d59b105b9c40a9518bbe958058df52423b595d7d9a293b71f8fe214f7bf88 Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.361771 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.507870 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-additional-scripts\") pod \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.507942 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69n4j\" (UniqueName: \"kubernetes.io/projected/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-kube-api-access-69n4j\") pod \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.508097 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-scripts\") pod \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.509267 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-scripts" (OuterVolumeSpecName: "scripts") pod "1727efd3-5377-4922-a7cb-4b3e9b0eabdb" (UID: "1727efd3-5377-4922-a7cb-4b3e9b0eabdb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.509295 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run-ovn\") pod \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.509504 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1727efd3-5377-4922-a7cb-4b3e9b0eabdb" (UID: "1727efd3-5377-4922-a7cb-4b3e9b0eabdb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.509670 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run" (OuterVolumeSpecName: "var-run") pod "1727efd3-5377-4922-a7cb-4b3e9b0eabdb" (UID: "1727efd3-5377-4922-a7cb-4b3e9b0eabdb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.509931 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run\") pod \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.509983 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-log-ovn\") pod \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\" (UID: \"1727efd3-5377-4922-a7cb-4b3e9b0eabdb\") " Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.510461 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1727efd3-5377-4922-a7cb-4b3e9b0eabdb" (UID: "1727efd3-5377-4922-a7cb-4b3e9b0eabdb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.510611 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1727efd3-5377-4922-a7cb-4b3e9b0eabdb" (UID: "1727efd3-5377-4922-a7cb-4b3e9b0eabdb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.511250 4675 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.511275 4675 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.511291 4675 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.511304 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.511532 4675 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.513898 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-kube-api-access-69n4j" (OuterVolumeSpecName: "kube-api-access-69n4j") pod "1727efd3-5377-4922-a7cb-4b3e9b0eabdb" (UID: "1727efd3-5377-4922-a7cb-4b3e9b0eabdb"). InnerVolumeSpecName "kube-api-access-69n4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.613467 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69n4j\" (UniqueName: \"kubernetes.io/projected/1727efd3-5377-4922-a7cb-4b3e9b0eabdb-kube-api-access-69n4j\") on node \"crc\" DevicePath \"\"" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.881667 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"e10696de0ee291811acecc01ec4785276621c2da79131d25baca1bce4cd58dbd"} Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.882036 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"52f4b33d9fea0dab8f0b5cbb847e9bd80309c7059365e6007a1065400aaea4bc"} Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.882050 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"74a1ef5b-9639-489e-9dc7-2da01bffa6f2","Type":"ContainerStarted","Data":"03b11e11d6ba48e6999a545c6a303c92555f9e8d2ce4b4d0e84896455f2d6197"} Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.883525 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc-config-5cj4n" event={"ID":"1727efd3-5377-4922-a7cb-4b3e9b0eabdb","Type":"ContainerDied","Data":"b69a030444321404b42076478306cf0a37382c8ee90622a8f43e5ed923e94db3"} Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.883552 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b69a030444321404b42076478306cf0a37382c8ee90622a8f43e5ed923e94db3" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.883551 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-5cj4n" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.884792 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4bphc" event={"ID":"fd86c751-bf59-47e4-81c4-dfaeb157fa76","Type":"ContainerStarted","Data":"4adba8d72c0f9e8863bb604064b1ee3146319595fb9e1e8b88c3bd5825a45d9e"} Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.884817 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4bphc" event={"ID":"fd86c751-bf59-47e4-81c4-dfaeb157fa76","Type":"ContainerStarted","Data":"9e7d59b105b9c40a9518bbe958058df52423b595d7d9a293b71f8fe214f7bf88"} Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.951022 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.898294216 podStartE2EDuration="27.951004198s" podCreationTimestamp="2026-02-19 09:00:28 +0000 UTC" firstStartedPulling="2026-02-19 09:00:45.610185857 +0000 UTC m=+1027.237276115" lastFinishedPulling="2026-02-19 09:00:53.662895829 +0000 UTC m=+1035.289986097" observedRunningTime="2026-02-19 09:00:55.945375285 +0000 UTC m=+1037.572465563" watchObservedRunningTime="2026-02-19 09:00:55.951004198 +0000 UTC m=+1037.578094466" Feb 19 09:00:55 crc kubenswrapper[4675]: I0219 09:00:55.987883 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-4bphc" podStartSLOduration=1.987865485 podStartE2EDuration="1.987865485s" podCreationTimestamp="2026-02-19 09:00:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:00:55.984775487 +0000 UTC m=+1037.611865765" watchObservedRunningTime="2026-02-19 09:00:55.987865485 +0000 UTC m=+1037.614955753" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.088810 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vlnhc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.300206 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-2xqlc"] Feb 19 09:00:56 crc kubenswrapper[4675]: E0219 09:00:56.300577 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1727efd3-5377-4922-a7cb-4b3e9b0eabdb" containerName="ovn-config" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.300592 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1727efd3-5377-4922-a7cb-4b3e9b0eabdb" containerName="ovn-config" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.300766 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="1727efd3-5377-4922-a7cb-4b3e9b0eabdb" containerName="ovn-config" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.301600 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.311880 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.330130 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-2xqlc"] Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.428295 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.428435 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s98lq\" (UniqueName: \"kubernetes.io/projected/645f6dc7-c3b9-4595-89ce-eff316454756-kube-api-access-s98lq\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.428466 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-config\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.428497 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.428881 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.428976 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.474871 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vlnhc-config-5cj4n"] Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.491460 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vlnhc-config-5cj4n"] Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.531118 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-config\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.531189 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.531264 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.531289 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.531322 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.531386 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s98lq\" (UniqueName: \"kubernetes.io/projected/645f6dc7-c3b9-4595-89ce-eff316454756-kube-api-access-s98lq\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.532865 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-config\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.533525 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.534152 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.534806 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.535335 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.555934 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s98lq\" (UniqueName: \"kubernetes.io/projected/645f6dc7-c3b9-4595-89ce-eff316454756-kube-api-access-s98lq\") pod \"dnsmasq-dns-6d5b6d6b67-2xqlc\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.627885 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vlnhc-config-sjl4f"] Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.630245 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.632775 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.645516 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vlnhc-config-sjl4f"] Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.678120 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.736626 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-scripts\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.737032 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-log-ovn\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.737198 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run-ovn\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.737316 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-additional-scripts\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.737460 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.737594 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-447ww\" (UniqueName: \"kubernetes.io/projected/147d509d-ad97-4911-9512-f024c10e0e63-kube-api-access-447ww\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.841774 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-log-ovn\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.841828 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run-ovn\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.841850 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-additional-scripts\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.841908 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.841941 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-447ww\" (UniqueName: \"kubernetes.io/projected/147d509d-ad97-4911-9512-f024c10e0e63-kube-api-access-447ww\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.842008 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-scripts\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.842943 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.842955 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-log-ovn\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.843002 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run-ovn\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.844383 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-additional-scripts\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.844469 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-scripts\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.861827 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-447ww\" (UniqueName: \"kubernetes.io/projected/147d509d-ad97-4911-9512-f024c10e0e63-kube-api-access-447ww\") pod \"ovn-controller-vlnhc-config-sjl4f\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.907080 4675 generic.go:334] "Generic (PLEG): container finished" podID="fd86c751-bf59-47e4-81c4-dfaeb157fa76" containerID="4adba8d72c0f9e8863bb604064b1ee3146319595fb9e1e8b88c3bd5825a45d9e" exitCode=0 Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.907176 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4bphc" event={"ID":"fd86c751-bf59-47e4-81c4-dfaeb157fa76","Type":"ContainerDied","Data":"4adba8d72c0f9e8863bb604064b1ee3146319595fb9e1e8b88c3bd5825a45d9e"} Feb 19 09:00:56 crc kubenswrapper[4675]: I0219 09:00:56.950873 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:00:57 crc kubenswrapper[4675]: I0219 09:00:57.114024 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1727efd3-5377-4922-a7cb-4b3e9b0eabdb" path="/var/lib/kubelet/pods/1727efd3-5377-4922-a7cb-4b3e9b0eabdb/volumes" Feb 19 09:01:02 crc kubenswrapper[4675]: I0219 09:01:02.800855 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.098499 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vhgrp"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.099595 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.117160 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.118678 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vhgrp"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.220362 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03b37867-7bec-498a-8ef7-2de4403950e8-operator-scripts\") pod \"cinder-db-create-vhgrp\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.220424 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg6s2\" (UniqueName: \"kubernetes.io/projected/03b37867-7bec-498a-8ef7-2de4403950e8-kube-api-access-mg6s2\") pod \"cinder-db-create-vhgrp\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.226340 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-bec0-account-create-update-mdjhn"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.227323 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.233941 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.254354 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bec0-account-create-update-mdjhn"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.322207 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03b37867-7bec-498a-8ef7-2de4403950e8-operator-scripts\") pod \"cinder-db-create-vhgrp\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.322253 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg6s2\" (UniqueName: \"kubernetes.io/projected/03b37867-7bec-498a-8ef7-2de4403950e8-kube-api-access-mg6s2\") pod \"cinder-db-create-vhgrp\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.322322 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13feadc4-3e1d-495b-8177-11af75487059-operator-scripts\") pod \"cinder-bec0-account-create-update-mdjhn\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.322411 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt9lt\" (UniqueName: \"kubernetes.io/projected/13feadc4-3e1d-495b-8177-11af75487059-kube-api-access-bt9lt\") pod \"cinder-bec0-account-create-update-mdjhn\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.322968 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03b37867-7bec-498a-8ef7-2de4403950e8-operator-scripts\") pod \"cinder-db-create-vhgrp\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.342938 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg6s2\" (UniqueName: \"kubernetes.io/projected/03b37867-7bec-498a-8ef7-2de4403950e8-kube-api-access-mg6s2\") pod \"cinder-db-create-vhgrp\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.408296 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-r8mnm"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.411392 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.423445 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1078-account-create-update-d8w87"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.424271 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13feadc4-3e1d-495b-8177-11af75487059-operator-scripts\") pod \"cinder-bec0-account-create-update-mdjhn\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.424402 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt9lt\" (UniqueName: \"kubernetes.io/projected/13feadc4-3e1d-495b-8177-11af75487059-kube-api-access-bt9lt\") pod \"cinder-bec0-account-create-update-mdjhn\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.424520 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.425821 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13feadc4-3e1d-495b-8177-11af75487059-operator-scripts\") pod \"cinder-bec0-account-create-update-mdjhn\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.427207 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.429002 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.434606 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-r8mnm"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.490575 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt9lt\" (UniqueName: \"kubernetes.io/projected/13feadc4-3e1d-495b-8177-11af75487059-kube-api-access-bt9lt\") pod \"cinder-bec0-account-create-update-mdjhn\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.508539 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1078-account-create-update-d8w87"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.526835 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-operator-scripts\") pod \"barbican-db-create-r8mnm\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.526897 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fjtw\" (UniqueName: \"kubernetes.io/projected/c8f55242-ce77-4241-ac85-e27d12376939-kube-api-access-2fjtw\") pod \"barbican-1078-account-create-update-d8w87\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.526957 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8f55242-ce77-4241-ac85-e27d12376939-operator-scripts\") pod \"barbican-1078-account-create-update-d8w87\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.527033 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwx6l\" (UniqueName: \"kubernetes.io/projected/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-kube-api-access-kwx6l\") pod \"barbican-db-create-r8mnm\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.553706 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.556500 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-8hnwf"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.558271 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.574370 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8hnwf"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.628814 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-operator-scripts\") pod \"barbican-db-create-r8mnm\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.628864 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fjtw\" (UniqueName: \"kubernetes.io/projected/c8f55242-ce77-4241-ac85-e27d12376939-kube-api-access-2fjtw\") pod \"barbican-1078-account-create-update-d8w87\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.628906 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8f55242-ce77-4241-ac85-e27d12376939-operator-scripts\") pod \"barbican-1078-account-create-update-d8w87\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.628957 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwx6l\" (UniqueName: \"kubernetes.io/projected/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-kube-api-access-kwx6l\") pod \"barbican-db-create-r8mnm\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.629937 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-operator-scripts\") pod \"barbican-db-create-r8mnm\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.630633 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8f55242-ce77-4241-ac85-e27d12376939-operator-scripts\") pod \"barbican-1078-account-create-update-d8w87\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.646786 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwx6l\" (UniqueName: \"kubernetes.io/projected/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-kube-api-access-kwx6l\") pod \"barbican-db-create-r8mnm\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.660308 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fjtw\" (UniqueName: \"kubernetes.io/projected/c8f55242-ce77-4241-ac85-e27d12376939-kube-api-access-2fjtw\") pod \"barbican-1078-account-create-update-d8w87\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.710983 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-flrm8"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.715869 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.728930 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6kbgs" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.730068 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a454b94-9aec-4a5a-b92c-6039497d605e-operator-scripts\") pod \"neutron-db-create-8hnwf\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.730125 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.730139 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv22b\" (UniqueName: \"kubernetes.io/projected/0a454b94-9aec-4a5a-b92c-6039497d605e-kube-api-access-kv22b\") pod \"neutron-db-create-8hnwf\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.730308 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.730558 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.737586 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.739115 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-flrm8"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.751134 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.788709 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-06ee-account-create-update-zpdvd"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.790034 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.808332 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.817015 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-06ee-account-create-update-zpdvd"] Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.836629 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-combined-ca-bundle\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.836740 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv22b\" (UniqueName: \"kubernetes.io/projected/0a454b94-9aec-4a5a-b92c-6039497d605e-kube-api-access-kv22b\") pod \"neutron-db-create-8hnwf\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.836903 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkdtv\" (UniqueName: \"kubernetes.io/projected/16be0916-9aea-444d-bc47-6bd7f646cb26-kube-api-access-nkdtv\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.836931 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-config-data\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.836977 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a454b94-9aec-4a5a-b92c-6039497d605e-operator-scripts\") pod \"neutron-db-create-8hnwf\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.837764 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a454b94-9aec-4a5a-b92c-6039497d605e-operator-scripts\") pod \"neutron-db-create-8hnwf\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.854542 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv22b\" (UniqueName: \"kubernetes.io/projected/0a454b94-9aec-4a5a-b92c-6039497d605e-kube-api-access-kv22b\") pod \"neutron-db-create-8hnwf\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.897106 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.940903 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkdtv\" (UniqueName: \"kubernetes.io/projected/16be0916-9aea-444d-bc47-6bd7f646cb26-kube-api-access-nkdtv\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.940953 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-config-data\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.941015 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n5br\" (UniqueName: \"kubernetes.io/projected/37e4b782-e0a5-4b72-8707-2c361c510f90-kube-api-access-7n5br\") pod \"neutron-06ee-account-create-update-zpdvd\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.941038 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-combined-ca-bundle\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.941078 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37e4b782-e0a5-4b72-8707-2c361c510f90-operator-scripts\") pod \"neutron-06ee-account-create-update-zpdvd\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.945097 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-config-data\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.947059 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-combined-ca-bundle\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:03 crc kubenswrapper[4675]: I0219 09:01:03.960134 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkdtv\" (UniqueName: \"kubernetes.io/projected/16be0916-9aea-444d-bc47-6bd7f646cb26-kube-api-access-nkdtv\") pod \"keystone-db-sync-flrm8\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:04 crc kubenswrapper[4675]: I0219 09:01:04.042879 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n5br\" (UniqueName: \"kubernetes.io/projected/37e4b782-e0a5-4b72-8707-2c361c510f90-kube-api-access-7n5br\") pod \"neutron-06ee-account-create-update-zpdvd\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:04 crc kubenswrapper[4675]: I0219 09:01:04.042950 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37e4b782-e0a5-4b72-8707-2c361c510f90-operator-scripts\") pod \"neutron-06ee-account-create-update-zpdvd\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:04 crc kubenswrapper[4675]: I0219 09:01:04.043842 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37e4b782-e0a5-4b72-8707-2c361c510f90-operator-scripts\") pod \"neutron-06ee-account-create-update-zpdvd\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:04 crc kubenswrapper[4675]: I0219 09:01:04.060142 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n5br\" (UniqueName: \"kubernetes.io/projected/37e4b782-e0a5-4b72-8707-2c361c510f90-kube-api-access-7n5br\") pod \"neutron-06ee-account-create-update-zpdvd\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:04 crc kubenswrapper[4675]: I0219 09:01:04.060258 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:04 crc kubenswrapper[4675]: I0219 09:01:04.197303 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:06 crc kubenswrapper[4675]: E0219 09:01:06.256444 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Feb 19 09:01:06 crc kubenswrapper[4675]: E0219 09:01:06.258829 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cxm6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-g8ngb_openstack(3ee47cd5-e516-4da9-ac5f-40cbedfe103b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:01:06 crc kubenswrapper[4675]: E0219 09:01:06.260155 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-g8ngb" podUID="3ee47cd5-e516-4da9-ac5f-40cbedfe103b" Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.423460 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4bphc" Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.608144 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd86c751-bf59-47e4-81c4-dfaeb157fa76-operator-scripts\") pod \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.608268 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgjs5\" (UniqueName: \"kubernetes.io/projected/fd86c751-bf59-47e4-81c4-dfaeb157fa76-kube-api-access-tgjs5\") pod \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\" (UID: \"fd86c751-bf59-47e4-81c4-dfaeb157fa76\") " Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.609284 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd86c751-bf59-47e4-81c4-dfaeb157fa76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fd86c751-bf59-47e4-81c4-dfaeb157fa76" (UID: "fd86c751-bf59-47e4-81c4-dfaeb157fa76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.636367 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd86c751-bf59-47e4-81c4-dfaeb157fa76-kube-api-access-tgjs5" (OuterVolumeSpecName: "kube-api-access-tgjs5") pod "fd86c751-bf59-47e4-81c4-dfaeb157fa76" (UID: "fd86c751-bf59-47e4-81c4-dfaeb157fa76"). InnerVolumeSpecName "kube-api-access-tgjs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.716458 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd86c751-bf59-47e4-81c4-dfaeb157fa76-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.716499 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgjs5\" (UniqueName: \"kubernetes.io/projected/fd86c751-bf59-47e4-81c4-dfaeb157fa76-kube-api-access-tgjs5\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:06 crc kubenswrapper[4675]: I0219 09:01:06.847575 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1078-account-create-update-d8w87"] Feb 19 09:01:06 crc kubenswrapper[4675]: W0219 09:01:06.877212 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8f55242_ce77_4241_ac85_e27d12376939.slice/crio-e59cd5eca6a9fc747ea880286c25d8a5c7568b384da25612c753c272edcbea75 WatchSource:0}: Error finding container e59cd5eca6a9fc747ea880286c25d8a5c7568b384da25612c753c272edcbea75: Status 404 returned error can't find the container with id e59cd5eca6a9fc747ea880286c25d8a5c7568b384da25612c753c272edcbea75 Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.008452 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8hnwf"] Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.023474 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1078-account-create-update-d8w87" event={"ID":"c8f55242-ce77-4241-ac85-e27d12376939","Type":"ContainerStarted","Data":"e59cd5eca6a9fc747ea880286c25d8a5c7568b384da25612c753c272edcbea75"} Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.031188 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4bphc" Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.033421 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4bphc" event={"ID":"fd86c751-bf59-47e4-81c4-dfaeb157fa76","Type":"ContainerDied","Data":"9e7d59b105b9c40a9518bbe958058df52423b595d7d9a293b71f8fe214f7bf88"} Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.033460 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e7d59b105b9c40a9518bbe958058df52423b595d7d9a293b71f8fe214f7bf88" Feb 19 09:01:07 crc kubenswrapper[4675]: E0219 09:01:07.033887 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-g8ngb" podUID="3ee47cd5-e516-4da9-ac5f-40cbedfe103b" Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.088106 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-2xqlc"] Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.141211 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vlnhc-config-sjl4f"] Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.345982 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-06ee-account-create-update-zpdvd"] Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.356333 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-r8mnm"] Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.381248 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-flrm8"] Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.528197 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bec0-account-create-update-mdjhn"] Feb 19 09:01:07 crc kubenswrapper[4675]: I0219 09:01:07.550693 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vhgrp"] Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.042046 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-flrm8" event={"ID":"16be0916-9aea-444d-bc47-6bd7f646cb26","Type":"ContainerStarted","Data":"4ad6f5f089223749a9102f8095983457fe619ee5d0f546a8f87ccf9090736425"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.049196 4675 generic.go:334] "Generic (PLEG): container finished" podID="0a454b94-9aec-4a5a-b92c-6039497d605e" containerID="be26d0805a20aa60bddde3bd34f2760db312e2e2b17b5ad4482b4eb5257af821" exitCode=0 Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.049319 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8hnwf" event={"ID":"0a454b94-9aec-4a5a-b92c-6039497d605e","Type":"ContainerDied","Data":"be26d0805a20aa60bddde3bd34f2760db312e2e2b17b5ad4482b4eb5257af821"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.049352 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8hnwf" event={"ID":"0a454b94-9aec-4a5a-b92c-6039497d605e","Type":"ContainerStarted","Data":"7cf138561d9a127d0f912b9720f04f5d62087452be506bf0f81f824e929acd47"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.052233 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bec0-account-create-update-mdjhn" event={"ID":"13feadc4-3e1d-495b-8177-11af75487059","Type":"ContainerStarted","Data":"b7a9391a34920b4b0e42d0a5d9ffca8cd522ff8098ce387c13da90f27dacb4f5"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.052286 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bec0-account-create-update-mdjhn" event={"ID":"13feadc4-3e1d-495b-8177-11af75487059","Type":"ContainerStarted","Data":"f4876b84f5d42cb11a66a9242c8901d8c13c891ad5fbbe2b5245ab8fca7c180d"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.060036 4675 generic.go:334] "Generic (PLEG): container finished" podID="645f6dc7-c3b9-4595-89ce-eff316454756" containerID="25f975d840cb9091c25aec3d46da36d20c77174a609fa0e249c0fc4c7be63e00" exitCode=0 Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.060116 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" event={"ID":"645f6dc7-c3b9-4595-89ce-eff316454756","Type":"ContainerDied","Data":"25f975d840cb9091c25aec3d46da36d20c77174a609fa0e249c0fc4c7be63e00"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.060153 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" event={"ID":"645f6dc7-c3b9-4595-89ce-eff316454756","Type":"ContainerStarted","Data":"294f35be9ce0b2d6cfe8f8680993b83eb8e8f39a0b298a744ba7fd5fc5859198"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.070901 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-r8mnm" event={"ID":"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9","Type":"ContainerStarted","Data":"a6ae9629c316ccf48c4bd31750bf148da87a1a6b9a8d434ef56a12e6a6d8652c"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.070950 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-r8mnm" event={"ID":"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9","Type":"ContainerStarted","Data":"1425ed61bec588a5434dbf8e062c79f08d89a605a53b3b3d1ea856968a748707"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.073878 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-06ee-account-create-update-zpdvd" event={"ID":"37e4b782-e0a5-4b72-8707-2c361c510f90","Type":"ContainerStarted","Data":"46466b799428bac83beddb04de2ca14b4c4ef45edfd5734b60933526dfd13ae0"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.073906 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-06ee-account-create-update-zpdvd" event={"ID":"37e4b782-e0a5-4b72-8707-2c361c510f90","Type":"ContainerStarted","Data":"c3b366f6c7b64ebce971886bc2bf18be3534552cbc1e19cb7ed9fb6032b29377"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.076550 4675 generic.go:334] "Generic (PLEG): container finished" podID="c8f55242-ce77-4241-ac85-e27d12376939" containerID="d5f80da71accb19bbb513b98cde1cd78644946a01c371aeaf2f124e3ed6dc557" exitCode=0 Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.076599 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1078-account-create-update-d8w87" event={"ID":"c8f55242-ce77-4241-ac85-e27d12376939","Type":"ContainerDied","Data":"d5f80da71accb19bbb513b98cde1cd78644946a01c371aeaf2f124e3ed6dc557"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.078774 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc-config-sjl4f" event={"ID":"147d509d-ad97-4911-9512-f024c10e0e63","Type":"ContainerStarted","Data":"ea3815d81b1cf5ae40f7aec970ce31de56cb826043899ccd1394fda58b38191d"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.078802 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc-config-sjl4f" event={"ID":"147d509d-ad97-4911-9512-f024c10e0e63","Type":"ContainerStarted","Data":"633dc04276a528060f15d3194da3b2cedd3918c3fd2dbb4a5d759b25e48a1b82"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.084066 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vhgrp" event={"ID":"03b37867-7bec-498a-8ef7-2de4403950e8","Type":"ContainerStarted","Data":"97a09a59c6405338e32577c28c1578002a771bb6b35b58cf1fc40eb301c9a94d"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.084135 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vhgrp" event={"ID":"03b37867-7bec-498a-8ef7-2de4403950e8","Type":"ContainerStarted","Data":"d59a51a7684b86eac3135af90d44f0272ea6a671f3be3430a84383a7b833bacc"} Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.087528 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-bec0-account-create-update-mdjhn" podStartSLOduration=5.087507177 podStartE2EDuration="5.087507177s" podCreationTimestamp="2026-02-19 09:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:08.082105263 +0000 UTC m=+1049.709195531" watchObservedRunningTime="2026-02-19 09:01:08.087507177 +0000 UTC m=+1049.714597445" Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.126394 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-06ee-account-create-update-zpdvd" podStartSLOduration=5.126362272 podStartE2EDuration="5.126362272s" podCreationTimestamp="2026-02-19 09:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:08.114828155 +0000 UTC m=+1049.741918423" watchObservedRunningTime="2026-02-19 09:01:08.126362272 +0000 UTC m=+1049.753452540" Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.137736 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-vhgrp" podStartSLOduration=5.137713735 podStartE2EDuration="5.137713735s" podCreationTimestamp="2026-02-19 09:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:08.133524823 +0000 UTC m=+1049.760615091" watchObservedRunningTime="2026-02-19 09:01:08.137713735 +0000 UTC m=+1049.764804003" Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.187435 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-r8mnm" podStartSLOduration=5.187402729 podStartE2EDuration="5.187402729s" podCreationTimestamp="2026-02-19 09:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:08.17882749 +0000 UTC m=+1049.805917778" watchObservedRunningTime="2026-02-19 09:01:08.187402729 +0000 UTC m=+1049.814492997" Feb 19 09:01:08 crc kubenswrapper[4675]: I0219 09:01:08.209188 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vlnhc-config-sjl4f" podStartSLOduration=12.209162209 podStartE2EDuration="12.209162209s" podCreationTimestamp="2026-02-19 09:00:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:08.196686256 +0000 UTC m=+1049.823776524" watchObservedRunningTime="2026-02-19 09:01:08.209162209 +0000 UTC m=+1049.836252497" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.099115 4675 generic.go:334] "Generic (PLEG): container finished" podID="a907b5a0-9709-4bc4-8b3a-7a364bcf66a9" containerID="a6ae9629c316ccf48c4bd31750bf148da87a1a6b9a8d434ef56a12e6a6d8652c" exitCode=0 Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.099184 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-r8mnm" event={"ID":"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9","Type":"ContainerDied","Data":"a6ae9629c316ccf48c4bd31750bf148da87a1a6b9a8d434ef56a12e6a6d8652c"} Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.101215 4675 generic.go:334] "Generic (PLEG): container finished" podID="37e4b782-e0a5-4b72-8707-2c361c510f90" containerID="46466b799428bac83beddb04de2ca14b4c4ef45edfd5734b60933526dfd13ae0" exitCode=0 Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.101265 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-06ee-account-create-update-zpdvd" event={"ID":"37e4b782-e0a5-4b72-8707-2c361c510f90","Type":"ContainerDied","Data":"46466b799428bac83beddb04de2ca14b4c4ef45edfd5734b60933526dfd13ae0"} Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.105338 4675 generic.go:334] "Generic (PLEG): container finished" podID="147d509d-ad97-4911-9512-f024c10e0e63" containerID="ea3815d81b1cf5ae40f7aec970ce31de56cb826043899ccd1394fda58b38191d" exitCode=0 Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.107079 4675 generic.go:334] "Generic (PLEG): container finished" podID="03b37867-7bec-498a-8ef7-2de4403950e8" containerID="97a09a59c6405338e32577c28c1578002a771bb6b35b58cf1fc40eb301c9a94d" exitCode=0 Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.108763 4675 generic.go:334] "Generic (PLEG): container finished" podID="13feadc4-3e1d-495b-8177-11af75487059" containerID="b7a9391a34920b4b0e42d0a5d9ffca8cd522ff8098ce387c13da90f27dacb4f5" exitCode=0 Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.117990 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc-config-sjl4f" event={"ID":"147d509d-ad97-4911-9512-f024c10e0e63","Type":"ContainerDied","Data":"ea3815d81b1cf5ae40f7aec970ce31de56cb826043899ccd1394fda58b38191d"} Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.118161 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vhgrp" event={"ID":"03b37867-7bec-498a-8ef7-2de4403950e8","Type":"ContainerDied","Data":"97a09a59c6405338e32577c28c1578002a771bb6b35b58cf1fc40eb301c9a94d"} Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.118177 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bec0-account-create-update-mdjhn" event={"ID":"13feadc4-3e1d-495b-8177-11af75487059","Type":"ContainerDied","Data":"b7a9391a34920b4b0e42d0a5d9ffca8cd522ff8098ce387c13da90f27dacb4f5"} Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.118191 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" event={"ID":"645f6dc7-c3b9-4595-89ce-eff316454756","Type":"ContainerStarted","Data":"322a7518cdb040945e0d8ac2593531619491766439e54623fbf6105fd05ea3b8"} Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.227476 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" podStartSLOduration=13.227455197 podStartE2EDuration="13.227455197s" podCreationTimestamp="2026-02-19 09:00:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:09.220719967 +0000 UTC m=+1050.847810235" watchObservedRunningTime="2026-02-19 09:01:09.227455197 +0000 UTC m=+1050.854545465" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.526991 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.553179 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.677270 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fjtw\" (UniqueName: \"kubernetes.io/projected/c8f55242-ce77-4241-ac85-e27d12376939-kube-api-access-2fjtw\") pod \"c8f55242-ce77-4241-ac85-e27d12376939\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.677326 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv22b\" (UniqueName: \"kubernetes.io/projected/0a454b94-9aec-4a5a-b92c-6039497d605e-kube-api-access-kv22b\") pod \"0a454b94-9aec-4a5a-b92c-6039497d605e\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.677349 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a454b94-9aec-4a5a-b92c-6039497d605e-operator-scripts\") pod \"0a454b94-9aec-4a5a-b92c-6039497d605e\" (UID: \"0a454b94-9aec-4a5a-b92c-6039497d605e\") " Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.677554 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8f55242-ce77-4241-ac85-e27d12376939-operator-scripts\") pod \"c8f55242-ce77-4241-ac85-e27d12376939\" (UID: \"c8f55242-ce77-4241-ac85-e27d12376939\") " Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.678609 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8f55242-ce77-4241-ac85-e27d12376939-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8f55242-ce77-4241-ac85-e27d12376939" (UID: "c8f55242-ce77-4241-ac85-e27d12376939"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.679081 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a454b94-9aec-4a5a-b92c-6039497d605e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a454b94-9aec-4a5a-b92c-6039497d605e" (UID: "0a454b94-9aec-4a5a-b92c-6039497d605e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.686270 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8f55242-ce77-4241-ac85-e27d12376939-kube-api-access-2fjtw" (OuterVolumeSpecName: "kube-api-access-2fjtw") pod "c8f55242-ce77-4241-ac85-e27d12376939" (UID: "c8f55242-ce77-4241-ac85-e27d12376939"). InnerVolumeSpecName "kube-api-access-2fjtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.692943 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a454b94-9aec-4a5a-b92c-6039497d605e-kube-api-access-kv22b" (OuterVolumeSpecName: "kube-api-access-kv22b") pod "0a454b94-9aec-4a5a-b92c-6039497d605e" (UID: "0a454b94-9aec-4a5a-b92c-6039497d605e"). InnerVolumeSpecName "kube-api-access-kv22b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.780390 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8f55242-ce77-4241-ac85-e27d12376939-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.780490 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fjtw\" (UniqueName: \"kubernetes.io/projected/c8f55242-ce77-4241-ac85-e27d12376939-kube-api-access-2fjtw\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.780509 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv22b\" (UniqueName: \"kubernetes.io/projected/0a454b94-9aec-4a5a-b92c-6039497d605e-kube-api-access-kv22b\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:09 crc kubenswrapper[4675]: I0219 09:01:09.780522 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a454b94-9aec-4a5a-b92c-6039497d605e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:10 crc kubenswrapper[4675]: I0219 09:01:10.119722 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8hnwf" event={"ID":"0a454b94-9aec-4a5a-b92c-6039497d605e","Type":"ContainerDied","Data":"7cf138561d9a127d0f912b9720f04f5d62087452be506bf0f81f824e929acd47"} Feb 19 09:01:10 crc kubenswrapper[4675]: I0219 09:01:10.119784 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cf138561d9a127d0f912b9720f04f5d62087452be506bf0f81f824e929acd47" Feb 19 09:01:10 crc kubenswrapper[4675]: I0219 09:01:10.119806 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8hnwf" Feb 19 09:01:10 crc kubenswrapper[4675]: I0219 09:01:10.125308 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1078-account-create-update-d8w87" Feb 19 09:01:10 crc kubenswrapper[4675]: I0219 09:01:10.126376 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1078-account-create-update-d8w87" event={"ID":"c8f55242-ce77-4241-ac85-e27d12376939","Type":"ContainerDied","Data":"e59cd5eca6a9fc747ea880286c25d8a5c7568b384da25612c753c272edcbea75"} Feb 19 09:01:10 crc kubenswrapper[4675]: I0219 09:01:10.126423 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e59cd5eca6a9fc747ea880286c25d8a5c7568b384da25612c753c272edcbea75" Feb 19 09:01:10 crc kubenswrapper[4675]: I0219 09:01:10.127598 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:01:11 crc kubenswrapper[4675]: I0219 09:01:11.743162 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:01:11 crc kubenswrapper[4675]: I0219 09:01:11.743536 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:01:11 crc kubenswrapper[4675]: I0219 09:01:11.743576 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:01:11 crc kubenswrapper[4675]: I0219 09:01:11.744080 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca5833a715b96b50d8038acbecde68c582755da122b5dea38657e570ff140d31"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:01:11 crc kubenswrapper[4675]: I0219 09:01:11.744130 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://ca5833a715b96b50d8038acbecde68c582755da122b5dea38657e570ff140d31" gracePeriod=600 Feb 19 09:01:13 crc kubenswrapper[4675]: I0219 09:01:13.156535 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="ca5833a715b96b50d8038acbecde68c582755da122b5dea38657e570ff140d31" exitCode=0 Feb 19 09:01:13 crc kubenswrapper[4675]: I0219 09:01:13.156600 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"ca5833a715b96b50d8038acbecde68c582755da122b5dea38657e570ff140d31"} Feb 19 09:01:13 crc kubenswrapper[4675]: I0219 09:01:13.157059 4675 scope.go:117] "RemoveContainer" containerID="8ee4c969741bfa1ae0382fcd0dac3640f8d7616539fc52b4655b88e929fd6acf" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.168659 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vlnhc-config-sjl4f" event={"ID":"147d509d-ad97-4911-9512-f024c10e0e63","Type":"ContainerDied","Data":"633dc04276a528060f15d3194da3b2cedd3918c3fd2dbb4a5d759b25e48a1b82"} Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.168982 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="633dc04276a528060f15d3194da3b2cedd3918c3fd2dbb4a5d759b25e48a1b82" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.170583 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vhgrp" event={"ID":"03b37867-7bec-498a-8ef7-2de4403950e8","Type":"ContainerDied","Data":"d59a51a7684b86eac3135af90d44f0272ea6a671f3be3430a84383a7b833bacc"} Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.170606 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d59a51a7684b86eac3135af90d44f0272ea6a671f3be3430a84383a7b833bacc" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.171915 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bec0-account-create-update-mdjhn" event={"ID":"13feadc4-3e1d-495b-8177-11af75487059","Type":"ContainerDied","Data":"f4876b84f5d42cb11a66a9242c8901d8c13c891ad5fbbe2b5245ab8fca7c180d"} Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.172038 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4876b84f5d42cb11a66a9242c8901d8c13c891ad5fbbe2b5245ab8fca7c180d" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.174052 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-r8mnm" event={"ID":"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9","Type":"ContainerDied","Data":"1425ed61bec588a5434dbf8e062c79f08d89a605a53b3b3d1ea856968a748707"} Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.174083 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1425ed61bec588a5434dbf8e062c79f08d89a605a53b3b3d1ea856968a748707" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.175494 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-06ee-account-create-update-zpdvd" event={"ID":"37e4b782-e0a5-4b72-8707-2c361c510f90","Type":"ContainerDied","Data":"c3b366f6c7b64ebce971886bc2bf18be3534552cbc1e19cb7ed9fb6032b29377"} Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.175521 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3b366f6c7b64ebce971886bc2bf18be3534552cbc1e19cb7ed9fb6032b29377" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.214768 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.237539 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.242438 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.247446 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.255407 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364460 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13feadc4-3e1d-495b-8177-11af75487059-operator-scripts\") pod \"13feadc4-3e1d-495b-8177-11af75487059\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364515 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg6s2\" (UniqueName: \"kubernetes.io/projected/03b37867-7bec-498a-8ef7-2de4403950e8-kube-api-access-mg6s2\") pod \"03b37867-7bec-498a-8ef7-2de4403950e8\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364609 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt9lt\" (UniqueName: \"kubernetes.io/projected/13feadc4-3e1d-495b-8177-11af75487059-kube-api-access-bt9lt\") pod \"13feadc4-3e1d-495b-8177-11af75487059\" (UID: \"13feadc4-3e1d-495b-8177-11af75487059\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364632 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run-ovn\") pod \"147d509d-ad97-4911-9512-f024c10e0e63\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364670 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03b37867-7bec-498a-8ef7-2de4403950e8-operator-scripts\") pod \"03b37867-7bec-498a-8ef7-2de4403950e8\" (UID: \"03b37867-7bec-498a-8ef7-2de4403950e8\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364732 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37e4b782-e0a5-4b72-8707-2c361c510f90-operator-scripts\") pod \"37e4b782-e0a5-4b72-8707-2c361c510f90\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364770 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-scripts\") pod \"147d509d-ad97-4911-9512-f024c10e0e63\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364788 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-log-ovn\") pod \"147d509d-ad97-4911-9512-f024c10e0e63\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364807 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n5br\" (UniqueName: \"kubernetes.io/projected/37e4b782-e0a5-4b72-8707-2c361c510f90-kube-api-access-7n5br\") pod \"37e4b782-e0a5-4b72-8707-2c361c510f90\" (UID: \"37e4b782-e0a5-4b72-8707-2c361c510f90\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364865 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-447ww\" (UniqueName: \"kubernetes.io/projected/147d509d-ad97-4911-9512-f024c10e0e63-kube-api-access-447ww\") pod \"147d509d-ad97-4911-9512-f024c10e0e63\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364880 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-operator-scripts\") pod \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364915 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwx6l\" (UniqueName: \"kubernetes.io/projected/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-kube-api-access-kwx6l\") pod \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\" (UID: \"a907b5a0-9709-4bc4-8b3a-7a364bcf66a9\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.364971 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-additional-scripts\") pod \"147d509d-ad97-4911-9512-f024c10e0e63\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.365020 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run\") pod \"147d509d-ad97-4911-9512-f024c10e0e63\" (UID: \"147d509d-ad97-4911-9512-f024c10e0e63\") " Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.365385 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run" (OuterVolumeSpecName: "var-run") pod "147d509d-ad97-4911-9512-f024c10e0e63" (UID: "147d509d-ad97-4911-9512-f024c10e0e63"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.366891 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-scripts" (OuterVolumeSpecName: "scripts") pod "147d509d-ad97-4911-9512-f024c10e0e63" (UID: "147d509d-ad97-4911-9512-f024c10e0e63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.367014 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "147d509d-ad97-4911-9512-f024c10e0e63" (UID: "147d509d-ad97-4911-9512-f024c10e0e63"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.367038 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "147d509d-ad97-4911-9512-f024c10e0e63" (UID: "147d509d-ad97-4911-9512-f024c10e0e63"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.367462 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a907b5a0-9709-4bc4-8b3a-7a364bcf66a9" (UID: "a907b5a0-9709-4bc4-8b3a-7a364bcf66a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.367514 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03b37867-7bec-498a-8ef7-2de4403950e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "03b37867-7bec-498a-8ef7-2de4403950e8" (UID: "03b37867-7bec-498a-8ef7-2de4403950e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.367783 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37e4b782-e0a5-4b72-8707-2c361c510f90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37e4b782-e0a5-4b72-8707-2c361c510f90" (UID: "37e4b782-e0a5-4b72-8707-2c361c510f90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.368065 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "147d509d-ad97-4911-9512-f024c10e0e63" (UID: "147d509d-ad97-4911-9512-f024c10e0e63"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.381845 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13feadc4-3e1d-495b-8177-11af75487059-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "13feadc4-3e1d-495b-8177-11af75487059" (UID: "13feadc4-3e1d-495b-8177-11af75487059"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.392880 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147d509d-ad97-4911-9512-f024c10e0e63-kube-api-access-447ww" (OuterVolumeSpecName: "kube-api-access-447ww") pod "147d509d-ad97-4911-9512-f024c10e0e63" (UID: "147d509d-ad97-4911-9512-f024c10e0e63"). InnerVolumeSpecName "kube-api-access-447ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.392994 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-kube-api-access-kwx6l" (OuterVolumeSpecName: "kube-api-access-kwx6l") pod "a907b5a0-9709-4bc4-8b3a-7a364bcf66a9" (UID: "a907b5a0-9709-4bc4-8b3a-7a364bcf66a9"). InnerVolumeSpecName "kube-api-access-kwx6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.393440 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37e4b782-e0a5-4b72-8707-2c361c510f90-kube-api-access-7n5br" (OuterVolumeSpecName: "kube-api-access-7n5br") pod "37e4b782-e0a5-4b72-8707-2c361c510f90" (UID: "37e4b782-e0a5-4b72-8707-2c361c510f90"). InnerVolumeSpecName "kube-api-access-7n5br". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.393517 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03b37867-7bec-498a-8ef7-2de4403950e8-kube-api-access-mg6s2" (OuterVolumeSpecName: "kube-api-access-mg6s2") pod "03b37867-7bec-498a-8ef7-2de4403950e8" (UID: "03b37867-7bec-498a-8ef7-2de4403950e8"). InnerVolumeSpecName "kube-api-access-mg6s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.407527 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13feadc4-3e1d-495b-8177-11af75487059-kube-api-access-bt9lt" (OuterVolumeSpecName: "kube-api-access-bt9lt") pod "13feadc4-3e1d-495b-8177-11af75487059" (UID: "13feadc4-3e1d-495b-8177-11af75487059"). InnerVolumeSpecName "kube-api-access-bt9lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.466747 4675 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467158 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13feadc4-3e1d-495b-8177-11af75487059-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467174 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg6s2\" (UniqueName: \"kubernetes.io/projected/03b37867-7bec-498a-8ef7-2de4403950e8-kube-api-access-mg6s2\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467186 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt9lt\" (UniqueName: \"kubernetes.io/projected/13feadc4-3e1d-495b-8177-11af75487059-kube-api-access-bt9lt\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467198 4675 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467209 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03b37867-7bec-498a-8ef7-2de4403950e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467220 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37e4b782-e0a5-4b72-8707-2c361c510f90-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467233 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467243 4675 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/147d509d-ad97-4911-9512-f024c10e0e63-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467256 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n5br\" (UniqueName: \"kubernetes.io/projected/37e4b782-e0a5-4b72-8707-2c361c510f90-kube-api-access-7n5br\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467267 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-447ww\" (UniqueName: \"kubernetes.io/projected/147d509d-ad97-4911-9512-f024c10e0e63-kube-api-access-447ww\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467278 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467289 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwx6l\" (UniqueName: \"kubernetes.io/projected/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9-kube-api-access-kwx6l\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:14 crc kubenswrapper[4675]: I0219 09:01:14.467301 4675 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/147d509d-ad97-4911-9512-f024c10e0e63-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.188785 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-flrm8" event={"ID":"16be0916-9aea-444d-bc47-6bd7f646cb26","Type":"ContainerStarted","Data":"4c2dfb0bcbabe81d458551f6acc664fb9e6bf529f3c18d617d1ced73c0d8670a"} Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.193034 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bec0-account-create-update-mdjhn" Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.193056 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"29c6cf62bac0dde894372357f7b07e5a365b8cc260e995aa58cda63d08dcff6f"} Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.193146 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-r8mnm" Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.193196 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vhgrp" Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.193201 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vlnhc-config-sjl4f" Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.193238 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-06ee-account-create-update-zpdvd" Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.209613 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-flrm8" podStartSLOduration=5.569705987 podStartE2EDuration="12.209594491s" podCreationTimestamp="2026-02-19 09:01:03 +0000 UTC" firstStartedPulling="2026-02-19 09:01:07.409963711 +0000 UTC m=+1049.037053979" lastFinishedPulling="2026-02-19 09:01:14.049852215 +0000 UTC m=+1055.676942483" observedRunningTime="2026-02-19 09:01:15.2091762 +0000 UTC m=+1056.836266478" watchObservedRunningTime="2026-02-19 09:01:15.209594491 +0000 UTC m=+1056.836684759" Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.415572 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vlnhc-config-sjl4f"] Feb 19 09:01:15 crc kubenswrapper[4675]: I0219 09:01:15.421867 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vlnhc-config-sjl4f"] Feb 19 09:01:16 crc kubenswrapper[4675]: I0219 09:01:16.679927 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:01:16 crc kubenswrapper[4675]: I0219 09:01:16.740003 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4b59q"] Feb 19 09:01:16 crc kubenswrapper[4675]: I0219 09:01:16.740581 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" podUID="ade75766-b387-4393-b66b-5a09665137fb" containerName="dnsmasq-dns" containerID="cri-o://f452536d0b73e2b61a8adf745b19b101d7c5a43a6d7d8adb0eec3a4df9e9bdab" gracePeriod=10 Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.113389 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147d509d-ad97-4911-9512-f024c10e0e63" path="/var/lib/kubelet/pods/147d509d-ad97-4911-9512-f024c10e0e63/volumes" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.209854 4675 generic.go:334] "Generic (PLEG): container finished" podID="ade75766-b387-4393-b66b-5a09665137fb" containerID="f452536d0b73e2b61a8adf745b19b101d7c5a43a6d7d8adb0eec3a4df9e9bdab" exitCode=0 Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.209922 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" event={"ID":"ade75766-b387-4393-b66b-5a09665137fb","Type":"ContainerDied","Data":"f452536d0b73e2b61a8adf745b19b101d7c5a43a6d7d8adb0eec3a4df9e9bdab"} Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.209973 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" event={"ID":"ade75766-b387-4393-b66b-5a09665137fb","Type":"ContainerDied","Data":"a45d6e31321f8650e4268d01d4cc5bd18c6747e45c0e17975231f18a1d306bcb"} Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.210009 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a45d6e31321f8650e4268d01d4cc5bd18c6747e45c0e17975231f18a1d306bcb" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.211705 4675 generic.go:334] "Generic (PLEG): container finished" podID="16be0916-9aea-444d-bc47-6bd7f646cb26" containerID="4c2dfb0bcbabe81d458551f6acc664fb9e6bf529f3c18d617d1ced73c0d8670a" exitCode=0 Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.211735 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-flrm8" event={"ID":"16be0916-9aea-444d-bc47-6bd7f646cb26","Type":"ContainerDied","Data":"4c2dfb0bcbabe81d458551f6acc664fb9e6bf529f3c18d617d1ced73c0d8670a"} Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.242182 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.424074 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-config\") pod \"ade75766-b387-4393-b66b-5a09665137fb\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.424196 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-dns-svc\") pod \"ade75766-b387-4393-b66b-5a09665137fb\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.424228 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqstf\" (UniqueName: \"kubernetes.io/projected/ade75766-b387-4393-b66b-5a09665137fb-kube-api-access-lqstf\") pod \"ade75766-b387-4393-b66b-5a09665137fb\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.424324 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-sb\") pod \"ade75766-b387-4393-b66b-5a09665137fb\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.424381 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-nb\") pod \"ade75766-b387-4393-b66b-5a09665137fb\" (UID: \"ade75766-b387-4393-b66b-5a09665137fb\") " Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.439050 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ade75766-b387-4393-b66b-5a09665137fb-kube-api-access-lqstf" (OuterVolumeSpecName: "kube-api-access-lqstf") pod "ade75766-b387-4393-b66b-5a09665137fb" (UID: "ade75766-b387-4393-b66b-5a09665137fb"). InnerVolumeSpecName "kube-api-access-lqstf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.524395 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ade75766-b387-4393-b66b-5a09665137fb" (UID: "ade75766-b387-4393-b66b-5a09665137fb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.527361 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-config" (OuterVolumeSpecName: "config") pod "ade75766-b387-4393-b66b-5a09665137fb" (UID: "ade75766-b387-4393-b66b-5a09665137fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.528793 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.528821 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqstf\" (UniqueName: \"kubernetes.io/projected/ade75766-b387-4393-b66b-5a09665137fb-kube-api-access-lqstf\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.528832 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.547499 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ade75766-b387-4393-b66b-5a09665137fb" (UID: "ade75766-b387-4393-b66b-5a09665137fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.551283 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ade75766-b387-4393-b66b-5a09665137fb" (UID: "ade75766-b387-4393-b66b-5a09665137fb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.631256 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:17 crc kubenswrapper[4675]: I0219 09:01:17.631290 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ade75766-b387-4393-b66b-5a09665137fb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.218505 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-4b59q" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.254590 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4b59q"] Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.260839 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-4b59q"] Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.545064 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.649734 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-config-data\") pod \"16be0916-9aea-444d-bc47-6bd7f646cb26\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.649923 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkdtv\" (UniqueName: \"kubernetes.io/projected/16be0916-9aea-444d-bc47-6bd7f646cb26-kube-api-access-nkdtv\") pod \"16be0916-9aea-444d-bc47-6bd7f646cb26\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.649969 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-combined-ca-bundle\") pod \"16be0916-9aea-444d-bc47-6bd7f646cb26\" (UID: \"16be0916-9aea-444d-bc47-6bd7f646cb26\") " Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.655940 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16be0916-9aea-444d-bc47-6bd7f646cb26-kube-api-access-nkdtv" (OuterVolumeSpecName: "kube-api-access-nkdtv") pod "16be0916-9aea-444d-bc47-6bd7f646cb26" (UID: "16be0916-9aea-444d-bc47-6bd7f646cb26"). InnerVolumeSpecName "kube-api-access-nkdtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.677417 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16be0916-9aea-444d-bc47-6bd7f646cb26" (UID: "16be0916-9aea-444d-bc47-6bd7f646cb26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.696865 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-config-data" (OuterVolumeSpecName: "config-data") pod "16be0916-9aea-444d-bc47-6bd7f646cb26" (UID: "16be0916-9aea-444d-bc47-6bd7f646cb26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.752092 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.752138 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkdtv\" (UniqueName: \"kubernetes.io/projected/16be0916-9aea-444d-bc47-6bd7f646cb26-kube-api-access-nkdtv\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:18 crc kubenswrapper[4675]: I0219 09:01:18.752149 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16be0916-9aea-444d-bc47-6bd7f646cb26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.118984 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ade75766-b387-4393-b66b-5a09665137fb" path="/var/lib/kubelet/pods/ade75766-b387-4393-b66b-5a09665137fb/volumes" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.228321 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-flrm8" event={"ID":"16be0916-9aea-444d-bc47-6bd7f646cb26","Type":"ContainerDied","Data":"4ad6f5f089223749a9102f8095983457fe619ee5d0f546a8f87ccf9090736425"} Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.228364 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ad6f5f089223749a9102f8095983457fe619ee5d0f546a8f87ccf9090736425" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.228435 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-flrm8" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.436499 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-svczw"] Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.436878 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b37867-7bec-498a-8ef7-2de4403950e8" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.436902 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b37867-7bec-498a-8ef7-2de4403950e8" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.436913 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd86c751-bf59-47e4-81c4-dfaeb157fa76" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.436922 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd86c751-bf59-47e4-81c4-dfaeb157fa76" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.436937 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a454b94-9aec-4a5a-b92c-6039497d605e" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.436944 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a454b94-9aec-4a5a-b92c-6039497d605e" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.436957 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16be0916-9aea-444d-bc47-6bd7f646cb26" containerName="keystone-db-sync" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.436963 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="16be0916-9aea-444d-bc47-6bd7f646cb26" containerName="keystone-db-sync" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.436974 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13feadc4-3e1d-495b-8177-11af75487059" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.436980 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="13feadc4-3e1d-495b-8177-11af75487059" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.436991 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e4b782-e0a5-4b72-8707-2c361c510f90" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.436998 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e4b782-e0a5-4b72-8707-2c361c510f90" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.437008 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a907b5a0-9709-4bc4-8b3a-7a364bcf66a9" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437013 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="a907b5a0-9709-4bc4-8b3a-7a364bcf66a9" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.437024 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade75766-b387-4393-b66b-5a09665137fb" containerName="init" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437030 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade75766-b387-4393-b66b-5a09665137fb" containerName="init" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.437042 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147d509d-ad97-4911-9512-f024c10e0e63" containerName="ovn-config" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437047 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="147d509d-ad97-4911-9512-f024c10e0e63" containerName="ovn-config" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.437054 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade75766-b387-4393-b66b-5a09665137fb" containerName="dnsmasq-dns" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437059 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade75766-b387-4393-b66b-5a09665137fb" containerName="dnsmasq-dns" Feb 19 09:01:19 crc kubenswrapper[4675]: E0219 09:01:19.437065 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8f55242-ce77-4241-ac85-e27d12376939" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437070 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8f55242-ce77-4241-ac85-e27d12376939" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437251 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade75766-b387-4393-b66b-5a09665137fb" containerName="dnsmasq-dns" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437266 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="a907b5a0-9709-4bc4-8b3a-7a364bcf66a9" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437275 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="16be0916-9aea-444d-bc47-6bd7f646cb26" containerName="keystone-db-sync" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437284 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a454b94-9aec-4a5a-b92c-6039497d605e" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437296 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd86c751-bf59-47e4-81c4-dfaeb157fa76" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437303 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="13feadc4-3e1d-495b-8177-11af75487059" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437312 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e4b782-e0a5-4b72-8707-2c361c510f90" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437318 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="147d509d-ad97-4911-9512-f024c10e0e63" containerName="ovn-config" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437330 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="03b37867-7bec-498a-8ef7-2de4403950e8" containerName="mariadb-database-create" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.437339 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8f55242-ce77-4241-ac85-e27d12376939" containerName="mariadb-account-create-update" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.438188 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.461678 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-svczw"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.503492 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2b55h"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.504702 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.507108 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.511080 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.511372 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.511541 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6kbgs" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.511707 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.534572 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2b55h"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.569813 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npcw2\" (UniqueName: \"kubernetes.io/projected/138ce145-ada0-4d2e-8d59-f8b86b925047-kube-api-access-npcw2\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.569865 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.569886 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-combined-ca-bundle\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.569912 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-credential-keys\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.569945 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.569971 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.569997 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.570053 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2g5l\" (UniqueName: \"kubernetes.io/projected/0824f353-528c-4847-acfd-8f8635d6102f-kube-api-access-j2g5l\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.570071 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-scripts\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.570092 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-config-data\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.570111 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-fernet-keys\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.570137 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-config\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672320 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2g5l\" (UniqueName: \"kubernetes.io/projected/0824f353-528c-4847-acfd-8f8635d6102f-kube-api-access-j2g5l\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672703 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-scripts\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672732 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-config-data\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672755 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-fernet-keys\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672787 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-config\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672823 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npcw2\" (UniqueName: \"kubernetes.io/projected/138ce145-ada0-4d2e-8d59-f8b86b925047-kube-api-access-npcw2\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672843 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672858 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-combined-ca-bundle\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672881 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-credential-keys\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672903 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672925 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.672943 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.673863 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.676160 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.676749 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.677783 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-config\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.678957 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.679591 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6dfb48549c-lkjnx"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.682290 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-combined-ca-bundle\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.683238 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-fernet-keys\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.683312 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-scripts\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.684416 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.686031 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-credential-keys\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.686868 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.687608 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.687894 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.696059 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-7xfst" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.698295 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npcw2\" (UniqueName: \"kubernetes.io/projected/138ce145-ada0-4d2e-8d59-f8b86b925047-kube-api-access-npcw2\") pod \"dnsmasq-dns-6f8c45789f-svczw\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.699423 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2g5l\" (UniqueName: \"kubernetes.io/projected/0824f353-528c-4847-acfd-8f8635d6102f-kube-api-access-j2g5l\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.701234 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6dfb48549c-lkjnx"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.712603 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-config-data\") pod \"keystone-bootstrap-2b55h\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.765994 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.797706 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.799663 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.812297 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-wmmx8"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.814208 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.816860 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.819096 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.819252 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pgv4v" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.819423 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.828977 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.844029 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.847756 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-wzzsx"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.848863 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.865736 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.866924 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.867739 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fhphd" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.868034 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.877553 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-config-data\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.877649 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f421816-e18a-4955-ba7a-557ed5e91ad3-horizon-secret-key\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.877709 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-scripts\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.877733 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk9d7\" (UniqueName: \"kubernetes.io/projected/0f421816-e18a-4955-ba7a-557ed5e91ad3-kube-api-access-lk9d7\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.877750 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f421816-e18a-4955-ba7a-557ed5e91ad3-logs\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.890251 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-fb8697f9f-57dnz"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.904212 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.911139 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wmmx8"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.961463 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fb8697f9f-57dnz"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.968828 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wzzsx"] Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.980228 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54802af-9255-48e3-b648-c5b7f6ba62db-logs\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.980548 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.980637 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-config-data\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.980737 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-log-httpd\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.980816 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-run-httpd\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.980904 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-combined-ca-bundle\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981175 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-scripts\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981277 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-config-data\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981354 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-db-sync-config-data\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981448 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f421816-e18a-4955-ba7a-557ed5e91ad3-horizon-secret-key\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981558 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-scripts\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981654 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d78a6299-1816-4500-9e03-a004f52262f8-etc-machine-id\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981754 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl8hr\" (UniqueName: \"kubernetes.io/projected/ecee41ce-0a05-4ea5-8f91-81f760d1f093-kube-api-access-nl8hr\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981836 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-config-data\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981903 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-config\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.981986 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-scripts\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982063 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54802af-9255-48e3-b648-c5b7f6ba62db-horizon-secret-key\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982139 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk9d7\" (UniqueName: \"kubernetes.io/projected/0f421816-e18a-4955-ba7a-557ed5e91ad3-kube-api-access-lk9d7\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982208 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f421816-e18a-4955-ba7a-557ed5e91ad3-logs\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982276 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982367 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-combined-ca-bundle\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982435 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g8v9\" (UniqueName: \"kubernetes.io/projected/a54802af-9255-48e3-b648-c5b7f6ba62db-kube-api-access-5g8v9\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982527 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-scripts\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982613 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvdds\" (UniqueName: \"kubernetes.io/projected/d78a6299-1816-4500-9e03-a004f52262f8-kube-api-access-xvdds\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982715 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tdf7\" (UniqueName: \"kubernetes.io/projected/0648a197-2cca-4ce8-bc6e-7162fd388de8-kube-api-access-4tdf7\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.982797 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-config-data\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.984787 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-config-data\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.986892 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f421816-e18a-4955-ba7a-557ed5e91ad3-logs\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.987548 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-scripts\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:19 crc kubenswrapper[4675]: I0219 09:01:19.995443 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f421816-e18a-4955-ba7a-557ed5e91ad3-horizon-secret-key\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.003890 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-cbncs"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.005176 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.013093 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mqn6h" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.013434 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.085137 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-scripts\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.088392 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvdds\" (UniqueName: \"kubernetes.io/projected/d78a6299-1816-4500-9e03-a004f52262f8-kube-api-access-xvdds\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.088700 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tdf7\" (UniqueName: \"kubernetes.io/projected/0648a197-2cca-4ce8-bc6e-7162fd388de8-kube-api-access-4tdf7\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.088807 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54802af-9255-48e3-b648-c5b7f6ba62db-logs\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.088924 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089040 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-config-data\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089162 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-log-httpd\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089265 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-run-httpd\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089366 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-combined-ca-bundle\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089465 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-scripts\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089563 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-config-data\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089674 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-db-sync-config-data\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089810 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-scripts\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.089996 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d78a6299-1816-4500-9e03-a004f52262f8-etc-machine-id\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.090129 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl8hr\" (UniqueName: \"kubernetes.io/projected/ecee41ce-0a05-4ea5-8f91-81f760d1f093-kube-api-access-nl8hr\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.090255 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-config-data\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.090354 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-config\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.090462 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54802af-9255-48e3-b648-c5b7f6ba62db-horizon-secret-key\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.090572 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.090701 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-combined-ca-bundle\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.090833 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g8v9\" (UniqueName: \"kubernetes.io/projected/a54802af-9255-48e3-b648-c5b7f6ba62db-kube-api-access-5g8v9\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.092177 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54802af-9255-48e3-b648-c5b7f6ba62db-logs\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.093491 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-run-httpd\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.096439 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d78a6299-1816-4500-9e03-a004f52262f8-etc-machine-id\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.142476 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.143455 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk9d7\" (UniqueName: \"kubernetes.io/projected/0f421816-e18a-4955-ba7a-557ed5e91ad3-kube-api-access-lk9d7\") pod \"horizon-6dfb48549c-lkjnx\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.144206 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-config\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.145081 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-db-sync-config-data\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.173330 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-config-data\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.175545 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-scripts\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.177612 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-log-httpd\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.188704 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-config-data\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.193869 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-scripts\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.203153 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-combined-ca-bundle\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.218385 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-combined-ca-bundle\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.219595 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.223239 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-cbncs"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.227555 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-scripts\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.227628 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvdds\" (UniqueName: \"kubernetes.io/projected/d78a6299-1816-4500-9e03-a004f52262f8-kube-api-access-xvdds\") pod \"cinder-db-sync-wzzsx\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.229763 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tdf7\" (UniqueName: \"kubernetes.io/projected/0648a197-2cca-4ce8-bc6e-7162fd388de8-kube-api-access-4tdf7\") pod \"neutron-db-sync-wmmx8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.251656 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-config-data\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.262091 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z27dc\" (UniqueName: \"kubernetes.io/projected/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-kube-api-access-z27dc\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.262167 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-combined-ca-bundle\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.262208 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-db-sync-config-data\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.265028 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54802af-9255-48e3-b648-c5b7f6ba62db-horizon-secret-key\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.272867 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g8v9\" (UniqueName: \"kubernetes.io/projected/a54802af-9255-48e3-b648-c5b7f6ba62db-kube-api-access-5g8v9\") pod \"horizon-fb8697f9f-57dnz\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.301173 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.308825 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl8hr\" (UniqueName: \"kubernetes.io/projected/ecee41ce-0a05-4ea5-8f91-81f760d1f093-kube-api-access-nl8hr\") pod \"ceilometer-0\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.365355 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z27dc\" (UniqueName: \"kubernetes.io/projected/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-kube-api-access-z27dc\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.365979 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-combined-ca-bundle\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.366015 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-db-sync-config-data\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.367408 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.369981 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.382003 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-db-sync-config-data\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.382078 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-wz69f"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.384396 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.394334 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-combined-ca-bundle\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.396502 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.397170 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.399522 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d8w68" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.404739 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-svczw"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.441789 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wz69f"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.457069 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z27dc\" (UniqueName: \"kubernetes.io/projected/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-kube-api-access-z27dc\") pod \"barbican-db-sync-cbncs\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.463153 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-pj9fm"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.468353 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824b7d51-57cb-459f-8876-9d88025ff60f-logs\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.468407 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-scripts\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.468439 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-combined-ca-bundle\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.468676 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-config-data\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.468746 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66sgh\" (UniqueName: \"kubernetes.io/projected/824b7d51-57cb-459f-8876-9d88025ff60f-kube-api-access-66sgh\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.471140 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.477458 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-pj9fm"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.519796 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572325 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-config-data\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572392 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7m8c\" (UniqueName: \"kubernetes.io/projected/e69df799-9ba4-4336-9f8f-dc08facd7c1b-kube-api-access-c7m8c\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572425 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66sgh\" (UniqueName: \"kubernetes.io/projected/824b7d51-57cb-459f-8876-9d88025ff60f-kube-api-access-66sgh\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572632 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572718 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572815 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572929 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824b7d51-57cb-459f-8876-9d88025ff60f-logs\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572957 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.572983 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-scripts\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.573009 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-combined-ca-bundle\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.573058 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-config\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.576030 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.578327 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824b7d51-57cb-459f-8876-9d88025ff60f-logs\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.581463 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-combined-ca-bundle\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.581599 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-scripts\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.582012 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-config-data\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.601869 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbncs" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.606334 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66sgh\" (UniqueName: \"kubernetes.io/projected/824b7d51-57cb-459f-8876-9d88025ff60f-kube-api-access-66sgh\") pod \"placement-db-sync-wz69f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.684992 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-config\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.685144 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7m8c\" (UniqueName: \"kubernetes.io/projected/e69df799-9ba4-4336-9f8f-dc08facd7c1b-kube-api-access-c7m8c\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.685200 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.685312 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.685408 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.685547 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.687523 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.692998 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.693865 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.694042 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-config\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.695299 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.709748 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7m8c\" (UniqueName: \"kubernetes.io/projected/e69df799-9ba4-4336-9f8f-dc08facd7c1b-kube-api-access-c7m8c\") pod \"dnsmasq-dns-fcfdd6f9f-pj9fm\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.744265 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wz69f" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.802464 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.816666 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2b55h"] Feb 19 09:01:20 crc kubenswrapper[4675]: I0219 09:01:20.826168 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-svczw"] Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.139840 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wmmx8"] Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.227302 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fb8697f9f-57dnz"] Feb 19 09:01:21 crc kubenswrapper[4675]: W0219 09:01:21.270106 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda54802af_9255_48e3_b648_c5b7f6ba62db.slice/crio-18ee12e498eb8cc4c8484db96903c0927bc35e81e1ac343b0f48103baeb7307d WatchSource:0}: Error finding container 18ee12e498eb8cc4c8484db96903c0927bc35e81e1ac343b0f48103baeb7307d: Status 404 returned error can't find the container with id 18ee12e498eb8cc4c8484db96903c0927bc35e81e1ac343b0f48103baeb7307d Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.272690 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6dfb48549c-lkjnx"] Feb 19 09:01:21 crc kubenswrapper[4675]: W0219 09:01:21.286425 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f421816_e18a_4955_ba7a_557ed5e91ad3.slice/crio-cb90b81a89ea792092d280f5c90c45369f10b6cfb73593e37fb24a9eecfb5241 WatchSource:0}: Error finding container cb90b81a89ea792092d280f5c90c45369f10b6cfb73593e37fb24a9eecfb5241: Status 404 returned error can't find the container with id cb90b81a89ea792092d280f5c90c45369f10b6cfb73593e37fb24a9eecfb5241 Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.310904 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wzzsx"] Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.319899 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wmmx8" event={"ID":"0648a197-2cca-4ce8-bc6e-7162fd388de8","Type":"ContainerStarted","Data":"394ca908569a62d6c05fb9c03e0f82e36977e9947adfdbadd857f58b26f88c64"} Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.324855 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-svczw" event={"ID":"138ce145-ada0-4d2e-8d59-f8b86b925047","Type":"ContainerStarted","Data":"2b5172b53086b9527d70b9832ea7362e4f2a9ce346e22270380273b18a04a8c2"} Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.327367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fb8697f9f-57dnz" event={"ID":"a54802af-9255-48e3-b648-c5b7f6ba62db","Type":"ContainerStarted","Data":"18ee12e498eb8cc4c8484db96903c0927bc35e81e1ac343b0f48103baeb7307d"} Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.347233 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b55h" event={"ID":"0824f353-528c-4847-acfd-8f8635d6102f","Type":"ContainerStarted","Data":"0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d"} Feb 19 09:01:21 crc kubenswrapper[4675]: W0219 09:01:21.376657 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd78a6299_1816_4500_9e03_a004f52262f8.slice/crio-81a65ebb8419dc93179494ed3fb0e97297ea368e793396b2ea5f51ea3c4ad701 WatchSource:0}: Error finding container 81a65ebb8419dc93179494ed3fb0e97297ea368e793396b2ea5f51ea3c4ad701: Status 404 returned error can't find the container with id 81a65ebb8419dc93179494ed3fb0e97297ea368e793396b2ea5f51ea3c4ad701 Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.447840 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wz69f"] Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.469615 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.652281 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-cbncs"] Feb 19 09:01:21 crc kubenswrapper[4675]: I0219 09:01:21.663716 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-pj9fm"] Feb 19 09:01:21 crc kubenswrapper[4675]: W0219 09:01:21.673548 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode69df799_9ba4_4336_9f8f_dc08facd7c1b.slice/crio-b48fca0047cfbc68f0870aee9afb17e650fc2be1eab34ef5e4eade4be73c1e63 WatchSource:0}: Error finding container b48fca0047cfbc68f0870aee9afb17e650fc2be1eab34ef5e4eade4be73c1e63: Status 404 returned error can't find the container with id b48fca0047cfbc68f0870aee9afb17e650fc2be1eab34ef5e4eade4be73c1e63 Feb 19 09:01:21 crc kubenswrapper[4675]: W0219 09:01:21.674084 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda9f4b70_4c48_4658_b2ae_b2d91b58f23c.slice/crio-6b6a8eb1f8e8ebdb184ed9ff853aa6c56caed54aba045c50eb0f8c7e50f35705 WatchSource:0}: Error finding container 6b6a8eb1f8e8ebdb184ed9ff853aa6c56caed54aba045c50eb0f8c7e50f35705: Status 404 returned error can't find the container with id 6b6a8eb1f8e8ebdb184ed9ff853aa6c56caed54aba045c50eb0f8c7e50f35705 Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.381186 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wzzsx" event={"ID":"d78a6299-1816-4500-9e03-a004f52262f8","Type":"ContainerStarted","Data":"81a65ebb8419dc93179494ed3fb0e97297ea368e793396b2ea5f51ea3c4ad701"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.392841 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbncs" event={"ID":"da9f4b70-4c48-4658-b2ae-b2d91b58f23c","Type":"ContainerStarted","Data":"6b6a8eb1f8e8ebdb184ed9ff853aa6c56caed54aba045c50eb0f8c7e50f35705"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.393507 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6dfb48549c-lkjnx"] Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.406485 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wmmx8" event={"ID":"0648a197-2cca-4ce8-bc6e-7162fd388de8","Type":"ContainerStarted","Data":"0a37609f47684f1831ddd07f069c7cf31f5b88a09a2bffbde4f3f4c461f9dca1"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.443433 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65bf577555-t6p57"] Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.445016 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.466337 4675 generic.go:334] "Generic (PLEG): container finished" podID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerID="48e1016435065fb0f5fc5e4aa7f2bc7700afdb6c45b8e62f4a0601755a9d2ad6" exitCode=0 Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.466423 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" event={"ID":"e69df799-9ba4-4336-9f8f-dc08facd7c1b","Type":"ContainerDied","Data":"48e1016435065fb0f5fc5e4aa7f2bc7700afdb6c45b8e62f4a0601755a9d2ad6"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.466449 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" event={"ID":"e69df799-9ba4-4336-9f8f-dc08facd7c1b","Type":"ContainerStarted","Data":"b48fca0047cfbc68f0870aee9afb17e650fc2be1eab34ef5e4eade4be73c1e63"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.515474 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.515524 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerStarted","Data":"234b9baa1d1bdbef293dd33d9a9502649b199a5f9dd800abec71534f272d17dc"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.535854 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65bf577555-t6p57"] Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.544441 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-wmmx8" podStartSLOduration=3.544419226 podStartE2EDuration="3.544419226s" podCreationTimestamp="2026-02-19 09:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:22.459386539 +0000 UTC m=+1064.086476807" watchObservedRunningTime="2026-02-19 09:01:22.544419226 +0000 UTC m=+1064.171509494" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.568874 4675 generic.go:334] "Generic (PLEG): container finished" podID="138ce145-ada0-4d2e-8d59-f8b86b925047" containerID="22ee10abbe8099511355bdf6c6454dca5148a53e3678654928932da3aed668ee" exitCode=0 Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.568997 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-svczw" event={"ID":"138ce145-ada0-4d2e-8d59-f8b86b925047","Type":"ContainerDied","Data":"22ee10abbe8099511355bdf6c6454dca5148a53e3678654928932da3aed668ee"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.599700 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wz69f" event={"ID":"824b7d51-57cb-459f-8876-9d88025ff60f","Type":"ContainerStarted","Data":"976645a53bb49fa4cc8175fc33f99f623e36ab02248875abc6a2e21f7933230f"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.626942 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g8ngb" event={"ID":"3ee47cd5-e516-4da9-ac5f-40cbedfe103b","Type":"ContainerStarted","Data":"4d7565a8ca901db4e12e439fdf1c78814cc7a9eba7b060a0b60df888676b8d14"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.647736 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2e38966-737c-49a6-800a-3293f1fd65a9-logs\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.647871 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2e38966-737c-49a6-800a-3293f1fd65a9-horizon-secret-key\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.647963 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqjhk\" (UniqueName: \"kubernetes.io/projected/d2e38966-737c-49a6-800a-3293f1fd65a9-kube-api-access-kqjhk\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.648109 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-scripts\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.648150 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-config-data\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.653143 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dfb48549c-lkjnx" event={"ID":"0f421816-e18a-4955-ba7a-557ed5e91ad3","Type":"ContainerStarted","Data":"cb90b81a89ea792092d280f5c90c45369f10b6cfb73593e37fb24a9eecfb5241"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.657690 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b55h" event={"ID":"0824f353-528c-4847-acfd-8f8635d6102f","Type":"ContainerStarted","Data":"323d24ff370c4234b2883d3742bba1699b5a106f42206d66ef5b910abcdb95a1"} Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.751911 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2e38966-737c-49a6-800a-3293f1fd65a9-horizon-secret-key\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.752041 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqjhk\" (UniqueName: \"kubernetes.io/projected/d2e38966-737c-49a6-800a-3293f1fd65a9-kube-api-access-kqjhk\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.752124 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-scripts\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.752182 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-config-data\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.752210 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2e38966-737c-49a6-800a-3293f1fd65a9-logs\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.752858 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2e38966-737c-49a6-800a-3293f1fd65a9-logs\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.773983 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-g8ngb" podStartSLOduration=4.816531416 podStartE2EDuration="36.773949083s" podCreationTimestamp="2026-02-19 09:00:46 +0000 UTC" firstStartedPulling="2026-02-19 09:00:47.641301366 +0000 UTC m=+1029.268391644" lastFinishedPulling="2026-02-19 09:01:19.598719043 +0000 UTC m=+1061.225809311" observedRunningTime="2026-02-19 09:01:22.667421494 +0000 UTC m=+1064.294511772" watchObservedRunningTime="2026-02-19 09:01:22.773949083 +0000 UTC m=+1064.401039351" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.780253 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2e38966-737c-49a6-800a-3293f1fd65a9-horizon-secret-key\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.781807 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-scripts\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.790041 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-config-data\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.798260 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2b55h" podStartSLOduration=3.79823054 podStartE2EDuration="3.79823054s" podCreationTimestamp="2026-02-19 09:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:22.695686247 +0000 UTC m=+1064.322776515" watchObservedRunningTime="2026-02-19 09:01:22.79823054 +0000 UTC m=+1064.425320808" Feb 19 09:01:22 crc kubenswrapper[4675]: I0219 09:01:22.827359 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqjhk\" (UniqueName: \"kubernetes.io/projected/d2e38966-737c-49a6-800a-3293f1fd65a9-kube-api-access-kqjhk\") pod \"horizon-65bf577555-t6p57\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:23 crc kubenswrapper[4675]: I0219 09:01:23.002414 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:23 crc kubenswrapper[4675]: I0219 09:01:23.199336 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.269897 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-swift-storage-0\") pod \"138ce145-ada0-4d2e-8d59-f8b86b925047\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.269961 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-svc\") pod \"138ce145-ada0-4d2e-8d59-f8b86b925047\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.269983 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-nb\") pod \"138ce145-ada0-4d2e-8d59-f8b86b925047\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.270016 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-sb\") pod \"138ce145-ada0-4d2e-8d59-f8b86b925047\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.270173 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-config\") pod \"138ce145-ada0-4d2e-8d59-f8b86b925047\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.270202 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npcw2\" (UniqueName: \"kubernetes.io/projected/138ce145-ada0-4d2e-8d59-f8b86b925047-kube-api-access-npcw2\") pod \"138ce145-ada0-4d2e-8d59-f8b86b925047\" (UID: \"138ce145-ada0-4d2e-8d59-f8b86b925047\") " Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.275876 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/138ce145-ada0-4d2e-8d59-f8b86b925047-kube-api-access-npcw2" (OuterVolumeSpecName: "kube-api-access-npcw2") pod "138ce145-ada0-4d2e-8d59-f8b86b925047" (UID: "138ce145-ada0-4d2e-8d59-f8b86b925047"). InnerVolumeSpecName "kube-api-access-npcw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.303869 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "138ce145-ada0-4d2e-8d59-f8b86b925047" (UID: "138ce145-ada0-4d2e-8d59-f8b86b925047"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.309183 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-config" (OuterVolumeSpecName: "config") pod "138ce145-ada0-4d2e-8d59-f8b86b925047" (UID: "138ce145-ada0-4d2e-8d59-f8b86b925047"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.309810 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "138ce145-ada0-4d2e-8d59-f8b86b925047" (UID: "138ce145-ada0-4d2e-8d59-f8b86b925047"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.311354 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "138ce145-ada0-4d2e-8d59-f8b86b925047" (UID: "138ce145-ada0-4d2e-8d59-f8b86b925047"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.319835 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "138ce145-ada0-4d2e-8d59-f8b86b925047" (UID: "138ce145-ada0-4d2e-8d59-f8b86b925047"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.373175 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.373259 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npcw2\" (UniqueName: \"kubernetes.io/projected/138ce145-ada0-4d2e-8d59-f8b86b925047-kube-api-access-npcw2\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.373271 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.373302 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.373311 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.373321 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/138ce145-ada0-4d2e-8d59-f8b86b925047-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.684592 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-svczw" event={"ID":"138ce145-ada0-4d2e-8d59-f8b86b925047","Type":"ContainerDied","Data":"2b5172b53086b9527d70b9832ea7362e4f2a9ce346e22270380273b18a04a8c2"} Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.684679 4675 scope.go:117] "RemoveContainer" containerID="22ee10abbe8099511355bdf6c6454dca5148a53e3678654928932da3aed668ee" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.684834 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-svczw" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.707732 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" event={"ID":"e69df799-9ba4-4336-9f8f-dc08facd7c1b","Type":"ContainerStarted","Data":"b48431ced98cfc828d5911429598b97b94ca60c2fb1c0a53790cdf84efa6180b"} Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.812027 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-svczw"] Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.822704 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-svczw"] Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:23.823738 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" podStartSLOduration=3.823702769 podStartE2EDuration="3.823702769s" podCreationTimestamp="2026-02-19 09:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:23.790172275 +0000 UTC m=+1065.417262543" watchObservedRunningTime="2026-02-19 09:01:23.823702769 +0000 UTC m=+1065.450793027" Feb 19 09:01:24 crc kubenswrapper[4675]: I0219 09:01:24.723554 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:25 crc kubenswrapper[4675]: I0219 09:01:25.091007 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65bf577555-t6p57"] Feb 19 09:01:25 crc kubenswrapper[4675]: W0219 09:01:25.138282 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2e38966_737c_49a6_800a_3293f1fd65a9.slice/crio-991b0cde90402a90d71ae6b88a533a27c71375fb74cad19cb493296e5486801e WatchSource:0}: Error finding container 991b0cde90402a90d71ae6b88a533a27c71375fb74cad19cb493296e5486801e: Status 404 returned error can't find the container with id 991b0cde90402a90d71ae6b88a533a27c71375fb74cad19cb493296e5486801e Feb 19 09:01:25 crc kubenswrapper[4675]: I0219 09:01:25.154270 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="138ce145-ada0-4d2e-8d59-f8b86b925047" path="/var/lib/kubelet/pods/138ce145-ada0-4d2e-8d59-f8b86b925047/volumes" Feb 19 09:01:25 crc kubenswrapper[4675]: I0219 09:01:25.742023 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bf577555-t6p57" event={"ID":"d2e38966-737c-49a6-800a-3293f1fd65a9","Type":"ContainerStarted","Data":"991b0cde90402a90d71ae6b88a533a27c71375fb74cad19cb493296e5486801e"} Feb 19 09:01:26 crc kubenswrapper[4675]: I0219 09:01:26.761683 4675 generic.go:334] "Generic (PLEG): container finished" podID="0824f353-528c-4847-acfd-8f8635d6102f" containerID="323d24ff370c4234b2883d3742bba1699b5a106f42206d66ef5b910abcdb95a1" exitCode=0 Feb 19 09:01:26 crc kubenswrapper[4675]: I0219 09:01:26.762096 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b55h" event={"ID":"0824f353-528c-4847-acfd-8f8635d6102f","Type":"ContainerDied","Data":"323d24ff370c4234b2883d3742bba1699b5a106f42206d66ef5b910abcdb95a1"} Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.683036 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fb8697f9f-57dnz"] Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.732324 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bbbd6c858-v6gv9"] Feb 19 09:01:29 crc kubenswrapper[4675]: E0219 09:01:29.732698 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138ce145-ada0-4d2e-8d59-f8b86b925047" containerName="init" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.732715 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="138ce145-ada0-4d2e-8d59-f8b86b925047" containerName="init" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.732905 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="138ce145-ada0-4d2e-8d59-f8b86b925047" containerName="init" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.733800 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.736463 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.751917 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bbbd6c858-v6gv9"] Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.822758 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65bf577555-t6p57"] Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.837627 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-combined-ca-bundle\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.837752 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-tls-certs\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.838055 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-secret-key\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.838234 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-config-data\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.838281 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb675\" (UniqueName: \"kubernetes.io/projected/d15338c6-aba7-4bec-a9c6-293e10903a7b-kube-api-access-qb675\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.838337 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15338c6-aba7-4bec-a9c6-293e10903a7b-logs\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.838378 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-scripts\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.864112 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64d55559f4-zwfgc"] Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.865517 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.893257 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d55559f4-zwfgc"] Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.939886 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-secret-key\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.939964 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-config-data\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.939990 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb675\" (UniqueName: \"kubernetes.io/projected/d15338c6-aba7-4bec-a9c6-293e10903a7b-kube-api-access-qb675\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.940013 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15338c6-aba7-4bec-a9c6-293e10903a7b-logs\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.940036 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-scripts\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.940068 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-combined-ca-bundle\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.940098 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-tls-certs\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.941454 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15338c6-aba7-4bec-a9c6-293e10903a7b-logs\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.942025 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-scripts\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.942042 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-config-data\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.948357 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-combined-ca-bundle\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.948623 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-tls-certs\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.949802 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-secret-key\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:29 crc kubenswrapper[4675]: I0219 09:01:29.972099 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb675\" (UniqueName: \"kubernetes.io/projected/d15338c6-aba7-4bec-a9c6-293e10903a7b-kube-api-access-qb675\") pod \"horizon-7bbbd6c858-v6gv9\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.041561 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18d90d12-7724-438d-8cd4-9fbe70c8ee89-logs\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.041671 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-horizon-secret-key\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.041703 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18d90d12-7724-438d-8cd4-9fbe70c8ee89-scripts\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.041735 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18d90d12-7724-438d-8cd4-9fbe70c8ee89-config-data\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.041914 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-horizon-tls-certs\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.041977 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-combined-ca-bundle\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.042237 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw26z\" (UniqueName: \"kubernetes.io/projected/18d90d12-7724-438d-8cd4-9fbe70c8ee89-kube-api-access-gw26z\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.056065 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.143586 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-horizon-secret-key\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.143713 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18d90d12-7724-438d-8cd4-9fbe70c8ee89-scripts\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.143774 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18d90d12-7724-438d-8cd4-9fbe70c8ee89-config-data\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.143804 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-horizon-tls-certs\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.144366 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-combined-ca-bundle\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.144504 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18d90d12-7724-438d-8cd4-9fbe70c8ee89-scripts\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.144524 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw26z\" (UniqueName: \"kubernetes.io/projected/18d90d12-7724-438d-8cd4-9fbe70c8ee89-kube-api-access-gw26z\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.144674 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18d90d12-7724-438d-8cd4-9fbe70c8ee89-logs\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.145134 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18d90d12-7724-438d-8cd4-9fbe70c8ee89-logs\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.145271 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18d90d12-7724-438d-8cd4-9fbe70c8ee89-config-data\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.148957 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-combined-ca-bundle\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.149997 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-horizon-tls-certs\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.162680 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18d90d12-7724-438d-8cd4-9fbe70c8ee89-horizon-secret-key\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.163352 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw26z\" (UniqueName: \"kubernetes.io/projected/18d90d12-7724-438d-8cd4-9fbe70c8ee89-kube-api-access-gw26z\") pod \"horizon-64d55559f4-zwfgc\" (UID: \"18d90d12-7724-438d-8cd4-9fbe70c8ee89\") " pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.189728 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.804667 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.864872 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-2xqlc"] Feb 19 09:01:30 crc kubenswrapper[4675]: I0219 09:01:30.865181 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" containerID="cri-o://322a7518cdb040945e0d8ac2593531619491766439e54623fbf6105fd05ea3b8" gracePeriod=10 Feb 19 09:01:31 crc kubenswrapper[4675]: I0219 09:01:31.678759 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 19 09:01:31 crc kubenswrapper[4675]: I0219 09:01:31.850050 4675 generic.go:334] "Generic (PLEG): container finished" podID="645f6dc7-c3b9-4595-89ce-eff316454756" containerID="322a7518cdb040945e0d8ac2593531619491766439e54623fbf6105fd05ea3b8" exitCode=0 Feb 19 09:01:31 crc kubenswrapper[4675]: I0219 09:01:31.850100 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" event={"ID":"645f6dc7-c3b9-4595-89ce-eff316454756","Type":"ContainerDied","Data":"322a7518cdb040945e0d8ac2593531619491766439e54623fbf6105fd05ea3b8"} Feb 19 09:01:36 crc kubenswrapper[4675]: I0219 09:01:36.678934 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 19 09:01:38 crc kubenswrapper[4675]: E0219 09:01:38.738500 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 19 09:01:38 crc kubenswrapper[4675]: E0219 09:01:38.739046 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65fh56ch5fbh564h58dh5d8h59h585h87h669h554h5b5hcfh5h5f4h7dh77h588hf7hc9hb8hd9h5dbh669hd5h59ch94h564h64ch566h4h598q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5g8v9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-fb8697f9f-57dnz_openstack(a54802af-9255-48e3-b648-c5b7f6ba62db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:01:38 crc kubenswrapper[4675]: E0219 09:01:38.743001 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-fb8697f9f-57dnz" podUID="a54802af-9255-48e3-b648-c5b7f6ba62db" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.799944 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.911657 4675 generic.go:334] "Generic (PLEG): container finished" podID="3ee47cd5-e516-4da9-ac5f-40cbedfe103b" containerID="4d7565a8ca901db4e12e439fdf1c78814cc7a9eba7b060a0b60df888676b8d14" exitCode=0 Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.911733 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g8ngb" event={"ID":"3ee47cd5-e516-4da9-ac5f-40cbedfe103b","Type":"ContainerDied","Data":"4d7565a8ca901db4e12e439fdf1c78814cc7a9eba7b060a0b60df888676b8d14"} Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.915854 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2b55h" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.916217 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2b55h" event={"ID":"0824f353-528c-4847-acfd-8f8635d6102f","Type":"ContainerDied","Data":"0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d"} Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.916257 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.922290 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-combined-ca-bundle\") pod \"0824f353-528c-4847-acfd-8f8635d6102f\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.922387 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-fernet-keys\") pod \"0824f353-528c-4847-acfd-8f8635d6102f\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.922424 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2g5l\" (UniqueName: \"kubernetes.io/projected/0824f353-528c-4847-acfd-8f8635d6102f-kube-api-access-j2g5l\") pod \"0824f353-528c-4847-acfd-8f8635d6102f\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.922479 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-credential-keys\") pod \"0824f353-528c-4847-acfd-8f8635d6102f\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.922576 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-config-data\") pod \"0824f353-528c-4847-acfd-8f8635d6102f\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.922723 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-scripts\") pod \"0824f353-528c-4847-acfd-8f8635d6102f\" (UID: \"0824f353-528c-4847-acfd-8f8635d6102f\") " Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.931790 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0824f353-528c-4847-acfd-8f8635d6102f-kube-api-access-j2g5l" (OuterVolumeSpecName: "kube-api-access-j2g5l") pod "0824f353-528c-4847-acfd-8f8635d6102f" (UID: "0824f353-528c-4847-acfd-8f8635d6102f"). InnerVolumeSpecName "kube-api-access-j2g5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.933905 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0824f353-528c-4847-acfd-8f8635d6102f" (UID: "0824f353-528c-4847-acfd-8f8635d6102f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.936200 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-scripts" (OuterVolumeSpecName: "scripts") pod "0824f353-528c-4847-acfd-8f8635d6102f" (UID: "0824f353-528c-4847-acfd-8f8635d6102f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.938270 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0824f353-528c-4847-acfd-8f8635d6102f" (UID: "0824f353-528c-4847-acfd-8f8635d6102f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.975824 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0824f353-528c-4847-acfd-8f8635d6102f" (UID: "0824f353-528c-4847-acfd-8f8635d6102f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:38 crc kubenswrapper[4675]: I0219 09:01:38.977168 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-config-data" (OuterVolumeSpecName: "config-data") pod "0824f353-528c-4847-acfd-8f8635d6102f" (UID: "0824f353-528c-4847-acfd-8f8635d6102f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.025962 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.025992 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.026001 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.026012 4675 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.026021 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2g5l\" (UniqueName: \"kubernetes.io/projected/0824f353-528c-4847-acfd-8f8635d6102f-kube-api-access-j2g5l\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.026029 4675 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0824f353-528c-4847-acfd-8f8635d6102f-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.917489 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2b55h"] Feb 19 09:01:39 crc kubenswrapper[4675]: I0219 09:01:39.926865 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2b55h"] Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.001227 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-c47wh"] Feb 19 09:01:40 crc kubenswrapper[4675]: E0219 09:01:40.002016 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0824f353-528c-4847-acfd-8f8635d6102f" containerName="keystone-bootstrap" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.002034 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0824f353-528c-4847-acfd-8f8635d6102f" containerName="keystone-bootstrap" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.002183 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0824f353-528c-4847-acfd-8f8635d6102f" containerName="keystone-bootstrap" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.002729 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.006508 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.006730 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.006853 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.007352 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6kbgs" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.007462 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.018923 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c47wh"] Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.156729 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-fernet-keys\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.156781 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-credential-keys\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.156805 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-config-data\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.157063 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-combined-ca-bundle\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.157313 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-scripts\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.157695 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-579mg\" (UniqueName: \"kubernetes.io/projected/e02b31ff-8f9a-4d65-b6fe-849630c89ede-kube-api-access-579mg\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.259669 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-combined-ca-bundle\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.259733 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-scripts\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.259780 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-579mg\" (UniqueName: \"kubernetes.io/projected/e02b31ff-8f9a-4d65-b6fe-849630c89ede-kube-api-access-579mg\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.259832 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-fernet-keys\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.259850 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-credential-keys\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.259870 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-config-data\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.265730 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-combined-ca-bundle\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.265761 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-credential-keys\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.266060 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-scripts\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.288747 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-579mg\" (UniqueName: \"kubernetes.io/projected/e02b31ff-8f9a-4d65-b6fe-849630c89ede-kube-api-access-579mg\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.289426 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-config-data\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.298035 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-fernet-keys\") pod \"keystone-bootstrap-c47wh\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:40 crc kubenswrapper[4675]: I0219 09:01:40.320352 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:01:41 crc kubenswrapper[4675]: I0219 09:01:41.114905 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0824f353-528c-4847-acfd-8f8635d6102f" path="/var/lib/kubelet/pods/0824f353-528c-4847-acfd-8f8635d6102f/volumes" Feb 19 09:01:41 crc kubenswrapper[4675]: E0219 09:01:41.447100 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Feb 19 09:01:41 crc kubenswrapper[4675]: E0219 09:01:41.447303 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-66sgh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-wz69f_openstack(824b7d51-57cb-459f-8876-9d88025ff60f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:01:41 crc kubenswrapper[4675]: E0219 09:01:41.449837 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-wz69f" podUID="824b7d51-57cb-459f-8876-9d88025ff60f" Feb 19 09:01:41 crc kubenswrapper[4675]: E0219 09:01:41.940523 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-wz69f" podUID="824b7d51-57cb-459f-8876-9d88025ff60f" Feb 19 09:01:43 crc kubenswrapper[4675]: E0219 09:01:43.710256 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0824f353_528c_4847_acfd_8f8635d6102f.slice/crio-0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d\": RecentStats: unable to find data in memory cache]" Feb 19 09:01:46 crc kubenswrapper[4675]: I0219 09:01:46.679675 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 19 09:01:46 crc kubenswrapper[4675]: I0219 09:01:46.681593 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:01:48 crc kubenswrapper[4675]: I0219 09:01:48.001963 4675 generic.go:334] "Generic (PLEG): container finished" podID="0648a197-2cca-4ce8-bc6e-7162fd388de8" containerID="0a37609f47684f1831ddd07f069c7cf31f5b88a09a2bffbde4f3f4c461f9dca1" exitCode=0 Feb 19 09:01:48 crc kubenswrapper[4675]: I0219 09:01:48.002042 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wmmx8" event={"ID":"0648a197-2cca-4ce8-bc6e-7162fd388de8","Type":"ContainerDied","Data":"0a37609f47684f1831ddd07f069c7cf31f5b88a09a2bffbde4f3f4c461f9dca1"} Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.012268 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" event={"ID":"645f6dc7-c3b9-4595-89ce-eff316454756","Type":"ContainerDied","Data":"294f35be9ce0b2d6cfe8f8680993b83eb8e8f39a0b298a744ba7fd5fc5859198"} Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.012554 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="294f35be9ce0b2d6cfe8f8680993b83eb8e8f39a0b298a744ba7fd5fc5859198" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.020553 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.132489 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-nb\") pod \"645f6dc7-c3b9-4595-89ce-eff316454756\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.132728 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-swift-storage-0\") pod \"645f6dc7-c3b9-4595-89ce-eff316454756\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.132891 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-config\") pod \"645f6dc7-c3b9-4595-89ce-eff316454756\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.132966 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-svc\") pod \"645f6dc7-c3b9-4595-89ce-eff316454756\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.133015 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-sb\") pod \"645f6dc7-c3b9-4595-89ce-eff316454756\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.133055 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s98lq\" (UniqueName: \"kubernetes.io/projected/645f6dc7-c3b9-4595-89ce-eff316454756-kube-api-access-s98lq\") pod \"645f6dc7-c3b9-4595-89ce-eff316454756\" (UID: \"645f6dc7-c3b9-4595-89ce-eff316454756\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.140932 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645f6dc7-c3b9-4595-89ce-eff316454756-kube-api-access-s98lq" (OuterVolumeSpecName: "kube-api-access-s98lq") pod "645f6dc7-c3b9-4595-89ce-eff316454756" (UID: "645f6dc7-c3b9-4595-89ce-eff316454756"). InnerVolumeSpecName "kube-api-access-s98lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.194119 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-config" (OuterVolumeSpecName: "config") pod "645f6dc7-c3b9-4595-89ce-eff316454756" (UID: "645f6dc7-c3b9-4595-89ce-eff316454756"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.210199 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "645f6dc7-c3b9-4595-89ce-eff316454756" (UID: "645f6dc7-c3b9-4595-89ce-eff316454756"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.212017 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "645f6dc7-c3b9-4595-89ce-eff316454756" (UID: "645f6dc7-c3b9-4595-89ce-eff316454756"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.216861 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "645f6dc7-c3b9-4595-89ce-eff316454756" (UID: "645f6dc7-c3b9-4595-89ce-eff316454756"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.217537 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "645f6dc7-c3b9-4595-89ce-eff316454756" (UID: "645f6dc7-c3b9-4595-89ce-eff316454756"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.239546 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.239590 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.239608 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.239618 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.239630 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s98lq\" (UniqueName: \"kubernetes.io/projected/645f6dc7-c3b9-4595-89ce-eff316454756-kube-api-access-s98lq\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.239662 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/645f6dc7-c3b9-4595-89ce-eff316454756-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: E0219 09:01:49.645109 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 19 09:01:49 crc kubenswrapper[4675]: E0219 09:01:49.645301 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z27dc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-cbncs_openstack(da9f4b70-4c48-4658-b2ae-b2d91b58f23c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:01:49 crc kubenswrapper[4675]: E0219 09:01:49.646406 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-cbncs" podUID="da9f4b70-4c48-4658-b2ae-b2d91b58f23c" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.654974 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g8ngb" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.658806 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748426 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54802af-9255-48e3-b648-c5b7f6ba62db-logs\") pod \"a54802af-9255-48e3-b648-c5b7f6ba62db\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748500 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxm6b\" (UniqueName: \"kubernetes.io/projected/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-kube-api-access-cxm6b\") pod \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748703 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g8v9\" (UniqueName: \"kubernetes.io/projected/a54802af-9255-48e3-b648-c5b7f6ba62db-kube-api-access-5g8v9\") pod \"a54802af-9255-48e3-b648-c5b7f6ba62db\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748733 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-config-data\") pod \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748751 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54802af-9255-48e3-b648-c5b7f6ba62db-horizon-secret-key\") pod \"a54802af-9255-48e3-b648-c5b7f6ba62db\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748775 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-combined-ca-bundle\") pod \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748814 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-scripts\") pod \"a54802af-9255-48e3-b648-c5b7f6ba62db\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748865 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-config-data\") pod \"a54802af-9255-48e3-b648-c5b7f6ba62db\" (UID: \"a54802af-9255-48e3-b648-c5b7f6ba62db\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748876 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a54802af-9255-48e3-b648-c5b7f6ba62db-logs" (OuterVolumeSpecName: "logs") pod "a54802af-9255-48e3-b648-c5b7f6ba62db" (UID: "a54802af-9255-48e3-b648-c5b7f6ba62db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.748965 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-db-sync-config-data\") pod \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\" (UID: \"3ee47cd5-e516-4da9-ac5f-40cbedfe103b\") " Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.749373 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54802af-9255-48e3-b648-c5b7f6ba62db-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.752071 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-scripts" (OuterVolumeSpecName: "scripts") pod "a54802af-9255-48e3-b648-c5b7f6ba62db" (UID: "a54802af-9255-48e3-b648-c5b7f6ba62db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.752717 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3ee47cd5-e516-4da9-ac5f-40cbedfe103b" (UID: "3ee47cd5-e516-4da9-ac5f-40cbedfe103b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.752882 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-config-data" (OuterVolumeSpecName: "config-data") pod "a54802af-9255-48e3-b648-c5b7f6ba62db" (UID: "a54802af-9255-48e3-b648-c5b7f6ba62db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.754979 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-kube-api-access-cxm6b" (OuterVolumeSpecName: "kube-api-access-cxm6b") pod "3ee47cd5-e516-4da9-ac5f-40cbedfe103b" (UID: "3ee47cd5-e516-4da9-ac5f-40cbedfe103b"). InnerVolumeSpecName "kube-api-access-cxm6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.755238 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54802af-9255-48e3-b648-c5b7f6ba62db-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a54802af-9255-48e3-b648-c5b7f6ba62db" (UID: "a54802af-9255-48e3-b648-c5b7f6ba62db"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.757362 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54802af-9255-48e3-b648-c5b7f6ba62db-kube-api-access-5g8v9" (OuterVolumeSpecName: "kube-api-access-5g8v9") pod "a54802af-9255-48e3-b648-c5b7f6ba62db" (UID: "a54802af-9255-48e3-b648-c5b7f6ba62db"). InnerVolumeSpecName "kube-api-access-5g8v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.772187 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ee47cd5-e516-4da9-ac5f-40cbedfe103b" (UID: "3ee47cd5-e516-4da9-ac5f-40cbedfe103b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.816022 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-config-data" (OuterVolumeSpecName: "config-data") pod "3ee47cd5-e516-4da9-ac5f-40cbedfe103b" (UID: "3ee47cd5-e516-4da9-ac5f-40cbedfe103b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851702 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g8v9\" (UniqueName: \"kubernetes.io/projected/a54802af-9255-48e3-b648-c5b7f6ba62db-kube-api-access-5g8v9\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851765 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851777 4675 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54802af-9255-48e3-b648-c5b7f6ba62db-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851787 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851798 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851807 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54802af-9255-48e3-b648-c5b7f6ba62db-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851816 4675 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:49 crc kubenswrapper[4675]: I0219 09:01:49.851824 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxm6b\" (UniqueName: \"kubernetes.io/projected/3ee47cd5-e516-4da9-ac5f-40cbedfe103b-kube-api-access-cxm6b\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.024031 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fb8697f9f-57dnz" event={"ID":"a54802af-9255-48e3-b648-c5b7f6ba62db","Type":"ContainerDied","Data":"18ee12e498eb8cc4c8484db96903c0927bc35e81e1ac343b0f48103baeb7307d"} Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.024108 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fb8697f9f-57dnz" Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.027046 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g8ngb" Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.037857 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g8ngb" event={"ID":"3ee47cd5-e516-4da9-ac5f-40cbedfe103b","Type":"ContainerDied","Data":"fdb9173f64fca44379a955831c660b3c255a53fdb9cb2398723e5b327f952032"} Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.037934 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdb9173f64fca44379a955831c660b3c255a53fdb9cb2398723e5b327f952032" Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.038066 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" Feb 19 09:01:50 crc kubenswrapper[4675]: E0219 09:01:50.048870 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-cbncs" podUID="da9f4b70-4c48-4658-b2ae-b2d91b58f23c" Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.083006 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-2xqlc"] Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.091721 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-2xqlc"] Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.121521 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fb8697f9f-57dnz"] Feb 19 09:01:50 crc kubenswrapper[4675]: I0219 09:01:50.131014 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-fb8697f9f-57dnz"] Feb 19 09:01:50 crc kubenswrapper[4675]: E0219 09:01:50.940898 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 19 09:01:50 crc kubenswrapper[4675]: E0219 09:01:50.941395 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xvdds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-wzzsx_openstack(d78a6299-1816-4500-9e03-a004f52262f8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:01:50 crc kubenswrapper[4675]: E0219 09:01:50.943380 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-wzzsx" podUID="d78a6299-1816-4500-9e03-a004f52262f8" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.072397 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wmmx8" event={"ID":"0648a197-2cca-4ce8-bc6e-7162fd388de8","Type":"ContainerDied","Data":"394ca908569a62d6c05fb9c03e0f82e36977e9947adfdbadd857f58b26f88c64"} Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.072438 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="394ca908569a62d6c05fb9c03e0f82e36977e9947adfdbadd857f58b26f88c64" Feb 19 09:01:51 crc kubenswrapper[4675]: E0219 09:01:51.075266 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-wzzsx" podUID="d78a6299-1816-4500-9e03-a004f52262f8" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.095102 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.158905 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" path="/var/lib/kubelet/pods/645f6dc7-c3b9-4595-89ce-eff316454756/volumes" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.177550 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a54802af-9255-48e3-b648-c5b7f6ba62db" path="/var/lib/kubelet/pods/a54802af-9255-48e3-b648-c5b7f6ba62db/volumes" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.189660 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tdf7\" (UniqueName: \"kubernetes.io/projected/0648a197-2cca-4ce8-bc6e-7162fd388de8-kube-api-access-4tdf7\") pod \"0648a197-2cca-4ce8-bc6e-7162fd388de8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.189836 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-config\") pod \"0648a197-2cca-4ce8-bc6e-7162fd388de8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.189892 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-combined-ca-bundle\") pod \"0648a197-2cca-4ce8-bc6e-7162fd388de8\" (UID: \"0648a197-2cca-4ce8-bc6e-7162fd388de8\") " Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.287299 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0648a197-2cca-4ce8-bc6e-7162fd388de8-kube-api-access-4tdf7" (OuterVolumeSpecName: "kube-api-access-4tdf7") pod "0648a197-2cca-4ce8-bc6e-7162fd388de8" (UID: "0648a197-2cca-4ce8-bc6e-7162fd388de8"). InnerVolumeSpecName "kube-api-access-4tdf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.293319 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tdf7\" (UniqueName: \"kubernetes.io/projected/0648a197-2cca-4ce8-bc6e-7162fd388de8-kube-api-access-4tdf7\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.351326 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-config" (OuterVolumeSpecName: "config") pod "0648a197-2cca-4ce8-bc6e-7162fd388de8" (UID: "0648a197-2cca-4ce8-bc6e-7162fd388de8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.399757 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.407958 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0648a197-2cca-4ce8-bc6e-7162fd388de8" (UID: "0648a197-2cca-4ce8-bc6e-7162fd388de8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.503146 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648a197-2cca-4ce8-bc6e-7162fd388de8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550150 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-76cp9"] Feb 19 09:01:51 crc kubenswrapper[4675]: E0219 09:01:51.550504 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550521 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" Feb 19 09:01:51 crc kubenswrapper[4675]: E0219 09:01:51.550534 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0648a197-2cca-4ce8-bc6e-7162fd388de8" containerName="neutron-db-sync" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550543 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0648a197-2cca-4ce8-bc6e-7162fd388de8" containerName="neutron-db-sync" Feb 19 09:01:51 crc kubenswrapper[4675]: E0219 09:01:51.550551 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="init" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550557 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="init" Feb 19 09:01:51 crc kubenswrapper[4675]: E0219 09:01:51.550583 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee47cd5-e516-4da9-ac5f-40cbedfe103b" containerName="glance-db-sync" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550591 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee47cd5-e516-4da9-ac5f-40cbedfe103b" containerName="glance-db-sync" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550767 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee47cd5-e516-4da9-ac5f-40cbedfe103b" containerName="glance-db-sync" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550780 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0648a197-2cca-4ce8-bc6e-7162fd388de8" containerName="neutron-db-sync" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.550787 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.551756 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-76cp9"] Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.551848 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.603963 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.604053 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.604085 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-config\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.604103 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.604124 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj6c5\" (UniqueName: \"kubernetes.io/projected/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-kube-api-access-jj6c5\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.604185 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.681837 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-2xqlc" podUID="645f6dc7-c3b9-4595-89ce-eff316454756" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.706717 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.706774 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.706830 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.706864 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-config\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.706884 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.706912 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj6c5\" (UniqueName: \"kubernetes.io/projected/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-kube-api-access-jj6c5\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.707849 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.707856 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.707998 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.708459 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-config\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.708555 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.731269 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj6c5\" (UniqueName: \"kubernetes.io/projected/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-kube-api-access-jj6c5\") pod \"dnsmasq-dns-57c957c4ff-76cp9\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.850574 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bbbd6c858-v6gv9"] Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.941597 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:51 crc kubenswrapper[4675]: I0219 09:01:51.993027 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c47wh"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.011473 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d55559f4-zwfgc"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.079227 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.081101 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.083323 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerStarted","Data":"6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.085228 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mswnr" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.085385 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.085812 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.086880 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.087197 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d55559f4-zwfgc" event={"ID":"18d90d12-7724-438d-8cd4-9fbe70c8ee89","Type":"ContainerStarted","Data":"ef2f2c52b61b4bdcbb6a8b2247231fc763d57e90304bf78c5d3721d12c0c969f"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.092288 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bf577555-t6p57" event={"ID":"d2e38966-737c-49a6-800a-3293f1fd65a9","Type":"ContainerStarted","Data":"80ec74317bd11d88956bcf2d002aa1462a757c1812cf26956b46d048b2b76569"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.092321 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bf577555-t6p57" event={"ID":"d2e38966-737c-49a6-800a-3293f1fd65a9","Type":"ContainerStarted","Data":"6c1fd05e0d490ce48f6728f5ab42948031b2d6da4ff8ab7aba93446c7a42a968"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.092447 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65bf577555-t6p57" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon-log" containerID="cri-o://6c1fd05e0d490ce48f6728f5ab42948031b2d6da4ff8ab7aba93446c7a42a968" gracePeriod=30 Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.092734 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65bf577555-t6p57" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon" containerID="cri-o://80ec74317bd11d88956bcf2d002aa1462a757c1812cf26956b46d048b2b76569" gracePeriod=30 Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.141158 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-65bf577555-t6p57" podStartSLOduration=4.318693601 podStartE2EDuration="30.141138312s" podCreationTimestamp="2026-02-19 09:01:22 +0000 UTC" firstStartedPulling="2026-02-19 09:01:25.156294943 +0000 UTC m=+1066.783385211" lastFinishedPulling="2026-02-19 09:01:50.978739654 +0000 UTC m=+1092.605829922" observedRunningTime="2026-02-19 09:01:52.129933604 +0000 UTC m=+1093.757023872" watchObservedRunningTime="2026-02-19 09:01:52.141138312 +0000 UTC m=+1093.768228580" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.152110 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bbbd6c858-v6gv9" event={"ID":"d15338c6-aba7-4bec-a9c6-293e10903a7b","Type":"ContainerStarted","Data":"b9b782a91e26d9785b05147cce47bc097bd4db31cf77a3e5cba4c3907d82264a"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.169092 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c47wh" event={"ID":"e02b31ff-8f9a-4d65-b6fe-849630c89ede","Type":"ContainerStarted","Data":"32bb29adc6cdac74a5542fd993141d0764de6128ab4b1c2b4821edb2095216a3"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.177038 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dfb48549c-lkjnx" event={"ID":"0f421816-e18a-4955-ba7a-557ed5e91ad3","Type":"ContainerStarted","Data":"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.177086 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dfb48549c-lkjnx" event={"ID":"0f421816-e18a-4955-ba7a-557ed5e91ad3","Type":"ContainerStarted","Data":"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5"} Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.177057 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wmmx8" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.177175 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6dfb48549c-lkjnx" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon-log" containerID="cri-o://40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5" gracePeriod=30 Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.177374 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6dfb48549c-lkjnx" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon" containerID="cri-o://006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad" gracePeriod=30 Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.215281 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6dfb48549c-lkjnx" podStartSLOduration=3.641235136 podStartE2EDuration="33.215262718s" podCreationTimestamp="2026-02-19 09:01:19 +0000 UTC" firstStartedPulling="2026-02-19 09:01:21.299011616 +0000 UTC m=+1062.926101894" lastFinishedPulling="2026-02-19 09:01:50.873039208 +0000 UTC m=+1092.500129476" observedRunningTime="2026-02-19 09:01:52.202796745 +0000 UTC m=+1093.829887013" watchObservedRunningTime="2026-02-19 09:01:52.215262718 +0000 UTC m=+1093.842352986" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.217239 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-scripts\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.217321 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-config-data\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.217352 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzjxw\" (UniqueName: \"kubernetes.io/projected/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-kube-api-access-tzjxw\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.217377 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-logs\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.217478 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.217528 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.217619 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.319140 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-scripts\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.319198 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-config-data\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.319236 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzjxw\" (UniqueName: \"kubernetes.io/projected/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-kube-api-access-tzjxw\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.319262 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-logs\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.319299 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.319370 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.319460 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.321490 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.321769 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-logs\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.322072 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.334405 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-scripts\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.334459 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.342303 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-config-data\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.356337 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzjxw\" (UniqueName: \"kubernetes.io/projected/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-kube-api-access-tzjxw\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.399589 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-76cp9"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.420175 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.422237 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.434824 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.455939 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.468254 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-vbwkt"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.471184 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.487889 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.520952 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-vbwkt"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.529487 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.529555 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.529576 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.529631 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.529727 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qdh8\" (UniqueName: \"kubernetes.io/projected/f067613d-aad1-426c-858b-2753c83c725e-kube-api-access-8qdh8\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.529781 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.529849 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.547162 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.584429 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-76cp9"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634037 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634593 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-config\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634613 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634652 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-565d4bc8d6-fvnsk"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634667 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qdh8\" (UniqueName: \"kubernetes.io/projected/f067613d-aad1-426c-858b-2753c83c725e-kube-api-access-8qdh8\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634886 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb48b\" (UniqueName: \"kubernetes.io/projected/1d0b3e27-6fb5-45da-b411-8e05d3119dff-kube-api-access-vb48b\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634939 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.634978 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.635019 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.635118 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.635156 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.635249 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.635292 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.635315 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.635994 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.636008 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.636325 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.638739 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.641001 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.641652 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.647223 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-565d4bc8d6-fvnsk"] Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.647944 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pgv4v" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.648124 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.652401 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.652573 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.664522 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qdh8\" (UniqueName: \"kubernetes.io/projected/f067613d-aad1-426c-858b-2753c83c725e-kube-api-access-8qdh8\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.672681 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.704493 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741036 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-httpd-config\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741142 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-config\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741170 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741220 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb48b\" (UniqueName: \"kubernetes.io/projected/1d0b3e27-6fb5-45da-b411-8e05d3119dff-kube-api-access-vb48b\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741242 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-config\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741270 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741311 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741396 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741452 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-ovndb-tls-certs\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741483 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-combined-ca-bundle\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.741506 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljfcf\" (UniqueName: \"kubernetes.io/projected/5bfdd470-3f89-435e-9781-2543a8d69a5b-kube-api-access-ljfcf\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.742442 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-config\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.743098 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.744309 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.745002 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.755724 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.768136 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.808487 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb48b\" (UniqueName: \"kubernetes.io/projected/1d0b3e27-6fb5-45da-b411-8e05d3119dff-kube-api-access-vb48b\") pod \"dnsmasq-dns-5ccc5c4795-vbwkt\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.814130 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.843635 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-config\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.844167 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-ovndb-tls-certs\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.844202 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-combined-ca-bundle\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.844219 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljfcf\" (UniqueName: \"kubernetes.io/projected/5bfdd470-3f89-435e-9781-2543a8d69a5b-kube-api-access-ljfcf\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.844252 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-httpd-config\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.849310 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-config\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.850348 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-ovndb-tls-certs\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.855437 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-combined-ca-bundle\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.857381 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-httpd-config\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.877352 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljfcf\" (UniqueName: \"kubernetes.io/projected/5bfdd470-3f89-435e-9781-2543a8d69a5b-kube-api-access-ljfcf\") pod \"neutron-565d4bc8d6-fvnsk\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:52 crc kubenswrapper[4675]: I0219 09:01:52.992091 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.003750 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.221592 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d55559f4-zwfgc" event={"ID":"18d90d12-7724-438d-8cd4-9fbe70c8ee89","Type":"ContainerStarted","Data":"d9db247a17990afadff5d764baae85007d71da183ed0d46a74593624f813a723"} Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.222210 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d55559f4-zwfgc" event={"ID":"18d90d12-7724-438d-8cd4-9fbe70c8ee89","Type":"ContainerStarted","Data":"082842e75aad99fa512f385b6e9ebef36d8860bd76cefcfa13d171b3564523b2"} Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.224245 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" event={"ID":"3e8ccb69-715a-4a12-9643-38a1a7fbde7d","Type":"ContainerStarted","Data":"314320a22088f19fb8f5913ab67d2cf90e6d7a3e486e267dea745bdb6a46c481"} Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.224499 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" podUID="3e8ccb69-715a-4a12-9643-38a1a7fbde7d" containerName="init" containerID="cri-o://d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a" gracePeriod=10 Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.321761 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bbbd6c858-v6gv9" event={"ID":"d15338c6-aba7-4bec-a9c6-293e10903a7b","Type":"ContainerStarted","Data":"0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2"} Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.321855 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bbbd6c858-v6gv9" event={"ID":"d15338c6-aba7-4bec-a9c6-293e10903a7b","Type":"ContainerStarted","Data":"56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff"} Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.334100 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c47wh" event={"ID":"e02b31ff-8f9a-4d65-b6fe-849630c89ede","Type":"ContainerStarted","Data":"1ce20a26d60177b6b285b56136a8da2029d771e37237d9234f57d0e3642d0c9f"} Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.383030 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.424667 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64d55559f4-zwfgc" podStartSLOduration=24.424628718 podStartE2EDuration="24.424628718s" podCreationTimestamp="2026-02-19 09:01:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:53.281255447 +0000 UTC m=+1094.908345725" watchObservedRunningTime="2026-02-19 09:01:53.424628718 +0000 UTC m=+1095.051718986" Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.440895 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bbbd6c858-v6gv9" podStartSLOduration=24.440875481 podStartE2EDuration="24.440875481s" podCreationTimestamp="2026-02-19 09:01:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:53.369345114 +0000 UTC m=+1094.996435392" watchObservedRunningTime="2026-02-19 09:01:53.440875481 +0000 UTC m=+1095.067965749" Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.458983 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-c47wh" podStartSLOduration=14.458957973 podStartE2EDuration="14.458957973s" podCreationTimestamp="2026-02-19 09:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:53.391380452 +0000 UTC m=+1095.018470730" watchObservedRunningTime="2026-02-19 09:01:53.458957973 +0000 UTC m=+1095.086048231" Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.546617 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-vbwkt"] Feb 19 09:01:53 crc kubenswrapper[4675]: W0219 09:01:53.672512 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf067613d_aad1_426c_858b_2753c83c725e.slice/crio-c5911a6655248dfe10c3e155c0d7e7d40a3a8cf482fcbc79a4edf5299edb85fd WatchSource:0}: Error finding container c5911a6655248dfe10c3e155c0d7e7d40a3a8cf482fcbc79a4edf5299edb85fd: Status 404 returned error can't find the container with id c5911a6655248dfe10c3e155c0d7e7d40a3a8cf482fcbc79a4edf5299edb85fd Feb 19 09:01:53 crc kubenswrapper[4675]: I0219 09:01:53.680468 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.059755 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-565d4bc8d6-fvnsk"] Feb 19 09:01:54 crc kubenswrapper[4675]: E0219 09:01:54.273302 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0824f353_528c_4847_acfd_8f8635d6102f.slice/crio-0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d\": RecentStats: unable to find data in memory cache]" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.346463 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.370979 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6ea0661b-0e96-4b12-af14-d5aab3d5edf9","Type":"ContainerStarted","Data":"74595bf7624e295e86af184965b45f0859350878a3e74b8ae6d757db4cdde1cd"} Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.407868 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-sb\") pod \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.407960 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-config\") pod \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.407990 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj6c5\" (UniqueName: \"kubernetes.io/projected/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-kube-api-access-jj6c5\") pod \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.408055 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-nb\") pod \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.408088 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-svc\") pod \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.408130 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-swift-storage-0\") pod \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\" (UID: \"3e8ccb69-715a-4a12-9643-38a1a7fbde7d\") " Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.408157 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" event={"ID":"3e8ccb69-715a-4a12-9643-38a1a7fbde7d","Type":"ContainerDied","Data":"d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a"} Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.408204 4675 scope.go:117] "RemoveContainer" containerID="d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.408399 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.408116 4675 generic.go:334] "Generic (PLEG): container finished" podID="3e8ccb69-715a-4a12-9643-38a1a7fbde7d" containerID="d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a" exitCode=0 Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.409726 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-76cp9" event={"ID":"3e8ccb69-715a-4a12-9643-38a1a7fbde7d","Type":"ContainerDied","Data":"314320a22088f19fb8f5913ab67d2cf90e6d7a3e486e267dea745bdb6a46c481"} Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.418351 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" event={"ID":"1d0b3e27-6fb5-45da-b411-8e05d3119dff","Type":"ContainerStarted","Data":"aec85350438df707cc5926e8c3094b9d6f0c92c6851206e3a40b20485044962d"} Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.435511 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f067613d-aad1-426c-858b-2753c83c725e","Type":"ContainerStarted","Data":"c5911a6655248dfe10c3e155c0d7e7d40a3a8cf482fcbc79a4edf5299edb85fd"} Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.446419 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-kube-api-access-jj6c5" (OuterVolumeSpecName: "kube-api-access-jj6c5") pod "3e8ccb69-715a-4a12-9643-38a1a7fbde7d" (UID: "3e8ccb69-715a-4a12-9643-38a1a7fbde7d"). InnerVolumeSpecName "kube-api-access-jj6c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.448566 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-565d4bc8d6-fvnsk" event={"ID":"5bfdd470-3f89-435e-9781-2543a8d69a5b","Type":"ContainerStarted","Data":"bd38504998f99efd6f6afc715e51da999b836e134a8db449e15e304a524307d2"} Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.465610 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3e8ccb69-715a-4a12-9643-38a1a7fbde7d" (UID: "3e8ccb69-715a-4a12-9643-38a1a7fbde7d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.483998 4675 scope.go:117] "RemoveContainer" containerID="d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a" Feb 19 09:01:54 crc kubenswrapper[4675]: E0219 09:01:54.484745 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a\": container with ID starting with d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a not found: ID does not exist" containerID="d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.484796 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a"} err="failed to get container status \"d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a\": rpc error: code = NotFound desc = could not find container \"d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a\": container with ID starting with d93643ccdd410410c2722b2012920bbd26f5880c4de6cf0b79e1ffcff24f535a not found: ID does not exist" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.493135 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e8ccb69-715a-4a12-9643-38a1a7fbde7d" (UID: "3e8ccb69-715a-4a12-9643-38a1a7fbde7d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.501928 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e8ccb69-715a-4a12-9643-38a1a7fbde7d" (UID: "3e8ccb69-715a-4a12-9643-38a1a7fbde7d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.503010 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-config" (OuterVolumeSpecName: "config") pod "3e8ccb69-715a-4a12-9643-38a1a7fbde7d" (UID: "3e8ccb69-715a-4a12-9643-38a1a7fbde7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.512404 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.512441 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj6c5\" (UniqueName: \"kubernetes.io/projected/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-kube-api-access-jj6c5\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.512452 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.512462 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.512471 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.519173 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e8ccb69-715a-4a12-9643-38a1a7fbde7d" (UID: "3e8ccb69-715a-4a12-9643-38a1a7fbde7d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.614358 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e8ccb69-715a-4a12-9643-38a1a7fbde7d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.815937 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-76cp9"] Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.835233 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-76cp9"] Feb 19 09:01:54 crc kubenswrapper[4675]: I0219 09:01:54.950057 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:55 crc kubenswrapper[4675]: I0219 09:01:55.073895 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:01:55 crc kubenswrapper[4675]: I0219 09:01:55.121968 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e8ccb69-715a-4a12-9643-38a1a7fbde7d" path="/var/lib/kubelet/pods/3e8ccb69-715a-4a12-9643-38a1a7fbde7d/volumes" Feb 19 09:01:55 crc kubenswrapper[4675]: I0219 09:01:55.464670 4675 generic.go:334] "Generic (PLEG): container finished" podID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerID="611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01" exitCode=0 Feb 19 09:01:55 crc kubenswrapper[4675]: I0219 09:01:55.464722 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" event={"ID":"1d0b3e27-6fb5-45da-b411-8e05d3119dff","Type":"ContainerDied","Data":"611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01"} Feb 19 09:01:55 crc kubenswrapper[4675]: I0219 09:01:55.473682 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f067613d-aad1-426c-858b-2753c83c725e","Type":"ContainerStarted","Data":"70230fe2d722bcf5790743c9b425ad6b79dbb41bea979ba0694852aecefabcfe"} Feb 19 09:01:55 crc kubenswrapper[4675]: I0219 09:01:55.475702 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6ea0661b-0e96-4b12-af14-d5aab3d5edf9","Type":"ContainerStarted","Data":"5f7ea628bf9d7ac1f8497fd680f2125eabb19a4867c44bbea201a039bb696a95"} Feb 19 09:01:56 crc kubenswrapper[4675]: I0219 09:01:56.515813 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-565d4bc8d6-fvnsk" event={"ID":"5bfdd470-3f89-435e-9781-2543a8d69a5b","Type":"ContainerStarted","Data":"681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571"} Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.531101 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" event={"ID":"1d0b3e27-6fb5-45da-b411-8e05d3119dff","Type":"ContainerStarted","Data":"b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508"} Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.532421 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.537593 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f067613d-aad1-426c-858b-2753c83c725e","Type":"ContainerStarted","Data":"4c3b476dae250f0b670a91f41a0b331b91a5660d3ca825eee266f1114ed77b48"} Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.538092 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-httpd" containerID="cri-o://4c3b476dae250f0b670a91f41a0b331b91a5660d3ca825eee266f1114ed77b48" gracePeriod=30 Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.538179 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-log" containerID="cri-o://70230fe2d722bcf5790743c9b425ad6b79dbb41bea979ba0694852aecefabcfe" gracePeriod=30 Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.542018 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-565d4bc8d6-fvnsk" event={"ID":"5bfdd470-3f89-435e-9781-2543a8d69a5b","Type":"ContainerStarted","Data":"0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a"} Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.543081 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.545750 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6ea0661b-0e96-4b12-af14-d5aab3d5edf9","Type":"ContainerStarted","Data":"bd78fac83ee615e73486f48d432ee8544674b7c3f75fce6420bf3413519cca4d"} Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.545911 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-log" containerID="cri-o://5f7ea628bf9d7ac1f8497fd680f2125eabb19a4867c44bbea201a039bb696a95" gracePeriod=30 Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.546029 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-httpd" containerID="cri-o://bd78fac83ee615e73486f48d432ee8544674b7c3f75fce6420bf3413519cca4d" gracePeriod=30 Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.558358 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerStarted","Data":"3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e"} Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.561442 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wz69f" event={"ID":"824b7d51-57cb-459f-8876-9d88025ff60f","Type":"ContainerStarted","Data":"9cf858bc93f6aefa286d61064d358262d8ef1f55263340fc054c72ac2ca4183e"} Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.572721 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" podStartSLOduration=5.572697034 podStartE2EDuration="5.572697034s" podCreationTimestamp="2026-02-19 09:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:57.555072515 +0000 UTC m=+1099.182162783" watchObservedRunningTime="2026-02-19 09:01:57.572697034 +0000 UTC m=+1099.199787292" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.583309 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.583290697 podStartE2EDuration="6.583290697s" podCreationTimestamp="2026-02-19 09:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:57.581464998 +0000 UTC m=+1099.208555286" watchObservedRunningTime="2026-02-19 09:01:57.583290697 +0000 UTC m=+1099.210380965" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.617821 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.617798136 podStartE2EDuration="6.617798136s" podCreationTimestamp="2026-02-19 09:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:57.608880949 +0000 UTC m=+1099.235971217" watchObservedRunningTime="2026-02-19 09:01:57.617798136 +0000 UTC m=+1099.244888404" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.663258 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-565d4bc8d6-fvnsk" podStartSLOduration=5.663233947 podStartE2EDuration="5.663233947s" podCreationTimestamp="2026-02-19 09:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:57.658330747 +0000 UTC m=+1099.285421015" watchObservedRunningTime="2026-02-19 09:01:57.663233947 +0000 UTC m=+1099.290324215" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.721468 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-wz69f" podStartSLOduration=3.304425741 podStartE2EDuration="37.721445139s" podCreationTimestamp="2026-02-19 09:01:20 +0000 UTC" firstStartedPulling="2026-02-19 09:01:21.509980908 +0000 UTC m=+1063.137071196" lastFinishedPulling="2026-02-19 09:01:55.927000326 +0000 UTC m=+1097.554090594" observedRunningTime="2026-02-19 09:01:57.700908631 +0000 UTC m=+1099.327998899" watchObservedRunningTime="2026-02-19 09:01:57.721445139 +0000 UTC m=+1099.348535407" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.788011 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7db6784559-bx2jh"] Feb 19 09:01:57 crc kubenswrapper[4675]: E0219 09:01:57.800329 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8ccb69-715a-4a12-9643-38a1a7fbde7d" containerName="init" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.800357 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8ccb69-715a-4a12-9643-38a1a7fbde7d" containerName="init" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.800572 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8ccb69-715a-4a12-9643-38a1a7fbde7d" containerName="init" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.802602 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.807939 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.808240 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.811944 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7db6784559-bx2jh"] Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.935953 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-public-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.936421 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-internal-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.936446 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-ovndb-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.936554 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-config\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.936577 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-combined-ca-bundle\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.936593 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b679c\" (UniqueName: \"kubernetes.io/projected/be374964-b557-48a9-b3db-68cf0e3d4a47-kube-api-access-b679c\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:57 crc kubenswrapper[4675]: I0219 09:01:57.936617 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-httpd-config\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.038559 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-config\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.038607 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-combined-ca-bundle\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.038629 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b679c\" (UniqueName: \"kubernetes.io/projected/be374964-b557-48a9-b3db-68cf0e3d4a47-kube-api-access-b679c\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.038670 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-httpd-config\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.038750 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-public-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.038796 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-internal-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.038814 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-ovndb-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.046903 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-httpd-config\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.047670 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-public-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.051774 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-ovndb-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.053887 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-config\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.054537 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-combined-ca-bundle\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.055237 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-internal-tls-certs\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.095821 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b679c\" (UniqueName: \"kubernetes.io/projected/be374964-b557-48a9-b3db-68cf0e3d4a47-kube-api-access-b679c\") pod \"neutron-7db6784559-bx2jh\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.154166 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.575607 4675 generic.go:334] "Generic (PLEG): container finished" podID="e02b31ff-8f9a-4d65-b6fe-849630c89ede" containerID="1ce20a26d60177b6b285b56136a8da2029d771e37237d9234f57d0e3642d0c9f" exitCode=0 Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.575726 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c47wh" event={"ID":"e02b31ff-8f9a-4d65-b6fe-849630c89ede","Type":"ContainerDied","Data":"1ce20a26d60177b6b285b56136a8da2029d771e37237d9234f57d0e3642d0c9f"} Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.603958 4675 generic.go:334] "Generic (PLEG): container finished" podID="f067613d-aad1-426c-858b-2753c83c725e" containerID="4c3b476dae250f0b670a91f41a0b331b91a5660d3ca825eee266f1114ed77b48" exitCode=0 Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.604055 4675 generic.go:334] "Generic (PLEG): container finished" podID="f067613d-aad1-426c-858b-2753c83c725e" containerID="70230fe2d722bcf5790743c9b425ad6b79dbb41bea979ba0694852aecefabcfe" exitCode=143 Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.604200 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f067613d-aad1-426c-858b-2753c83c725e","Type":"ContainerDied","Data":"4c3b476dae250f0b670a91f41a0b331b91a5660d3ca825eee266f1114ed77b48"} Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.604244 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f067613d-aad1-426c-858b-2753c83c725e","Type":"ContainerDied","Data":"70230fe2d722bcf5790743c9b425ad6b79dbb41bea979ba0694852aecefabcfe"} Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.623053 4675 generic.go:334] "Generic (PLEG): container finished" podID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerID="bd78fac83ee615e73486f48d432ee8544674b7c3f75fce6420bf3413519cca4d" exitCode=0 Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.623102 4675 generic.go:334] "Generic (PLEG): container finished" podID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerID="5f7ea628bf9d7ac1f8497fd680f2125eabb19a4867c44bbea201a039bb696a95" exitCode=143 Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.624123 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6ea0661b-0e96-4b12-af14-d5aab3d5edf9","Type":"ContainerDied","Data":"bd78fac83ee615e73486f48d432ee8544674b7c3f75fce6420bf3413519cca4d"} Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.624158 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6ea0661b-0e96-4b12-af14-d5aab3d5edf9","Type":"ContainerDied","Data":"5f7ea628bf9d7ac1f8497fd680f2125eabb19a4867c44bbea201a039bb696a95"} Feb 19 09:01:58 crc kubenswrapper[4675]: I0219 09:01:58.859830 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7db6784559-bx2jh"] Feb 19 09:01:58 crc kubenswrapper[4675]: W0219 09:01:58.872430 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe374964_b557_48a9_b3db_68cf0e3d4a47.slice/crio-5f1b56c4ce8e4faf94f514e437b648a56d325f8004ad928c609e9579136a88b8 WatchSource:0}: Error finding container 5f1b56c4ce8e4faf94f514e437b648a56d325f8004ad928c609e9579136a88b8: Status 404 returned error can't find the container with id 5f1b56c4ce8e4faf94f514e437b648a56d325f8004ad928c609e9579136a88b8 Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.064703 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.165167 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-combined-ca-bundle\") pod \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.203903 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ea0661b-0e96-4b12-af14-d5aab3d5edf9" (UID: "6ea0661b-0e96-4b12-af14-d5aab3d5edf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.268564 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-config-data\") pod \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.268966 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzjxw\" (UniqueName: \"kubernetes.io/projected/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-kube-api-access-tzjxw\") pod \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.269137 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-logs\") pod \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.269155 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-httpd-run\") pod \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.269219 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-scripts\") pod \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.269255 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\" (UID: \"6ea0661b-0e96-4b12-af14-d5aab3d5edf9\") " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.269746 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.270227 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6ea0661b-0e96-4b12-af14-d5aab3d5edf9" (UID: "6ea0661b-0e96-4b12-af14-d5aab3d5edf9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.270683 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-logs" (OuterVolumeSpecName: "logs") pod "6ea0661b-0e96-4b12-af14-d5aab3d5edf9" (UID: "6ea0661b-0e96-4b12-af14-d5aab3d5edf9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.274806 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "6ea0661b-0e96-4b12-af14-d5aab3d5edf9" (UID: "6ea0661b-0e96-4b12-af14-d5aab3d5edf9"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.278748 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-kube-api-access-tzjxw" (OuterVolumeSpecName: "kube-api-access-tzjxw") pod "6ea0661b-0e96-4b12-af14-d5aab3d5edf9" (UID: "6ea0661b-0e96-4b12-af14-d5aab3d5edf9"). InnerVolumeSpecName "kube-api-access-tzjxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.281571 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-scripts" (OuterVolumeSpecName: "scripts") pod "6ea0661b-0e96-4b12-af14-d5aab3d5edf9" (UID: "6ea0661b-0e96-4b12-af14-d5aab3d5edf9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.335012 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-config-data" (OuterVolumeSpecName: "config-data") pod "6ea0661b-0e96-4b12-af14-d5aab3d5edf9" (UID: "6ea0661b-0e96-4b12-af14-d5aab3d5edf9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.371143 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzjxw\" (UniqueName: \"kubernetes.io/projected/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-kube-api-access-tzjxw\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.371173 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.371185 4675 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.371195 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.371229 4675 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.371239 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea0661b-0e96-4b12-af14-d5aab3d5edf9-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.391984 4675 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.472688 4675 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.640621 4675 generic.go:334] "Generic (PLEG): container finished" podID="824b7d51-57cb-459f-8876-9d88025ff60f" containerID="9cf858bc93f6aefa286d61064d358262d8ef1f55263340fc054c72ac2ca4183e" exitCode=0 Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.640715 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wz69f" event={"ID":"824b7d51-57cb-459f-8876-9d88025ff60f","Type":"ContainerDied","Data":"9cf858bc93f6aefa286d61064d358262d8ef1f55263340fc054c72ac2ca4183e"} Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.647136 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6ea0661b-0e96-4b12-af14-d5aab3d5edf9","Type":"ContainerDied","Data":"74595bf7624e295e86af184965b45f0859350878a3e74b8ae6d757db4cdde1cd"} Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.647184 4675 scope.go:117] "RemoveContainer" containerID="bd78fac83ee615e73486f48d432ee8544674b7c3f75fce6420bf3413519cca4d" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.647307 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.657874 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db6784559-bx2jh" event={"ID":"be374964-b557-48a9-b3db-68cf0e3d4a47","Type":"ContainerStarted","Data":"aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02"} Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.657918 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db6784559-bx2jh" event={"ID":"be374964-b557-48a9-b3db-68cf0e3d4a47","Type":"ContainerStarted","Data":"bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28"} Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.657937 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.657949 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db6784559-bx2jh" event={"ID":"be374964-b557-48a9-b3db-68cf0e3d4a47","Type":"ContainerStarted","Data":"5f1b56c4ce8e4faf94f514e437b648a56d325f8004ad928c609e9579136a88b8"} Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.713586 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7db6784559-bx2jh" podStartSLOduration=2.713563358 podStartE2EDuration="2.713563358s" podCreationTimestamp="2026-02-19 09:01:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:01:59.71211585 +0000 UTC m=+1101.339206118" watchObservedRunningTime="2026-02-19 09:01:59.713563358 +0000 UTC m=+1101.340653626" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.788324 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.789323 4675 scope.go:117] "RemoveContainer" containerID="5f7ea628bf9d7ac1f8497fd680f2125eabb19a4867c44bbea201a039bb696a95" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.807858 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.851911 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:59 crc kubenswrapper[4675]: E0219 09:01:59.852332 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-httpd" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.852349 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-httpd" Feb 19 09:01:59 crc kubenswrapper[4675]: E0219 09:01:59.852378 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-log" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.852384 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-log" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.852560 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-httpd" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.852581 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" containerName="glance-log" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.853533 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.864125 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.864386 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.864821 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985034 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985106 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rztfj\" (UniqueName: \"kubernetes.io/projected/d2589e24-7bb6-49ac-a98c-55be590405df-kube-api-access-rztfj\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985372 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985424 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985446 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985468 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-logs\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985486 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:01:59 crc kubenswrapper[4675]: I0219 09:01:59.985555 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.057793 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.057880 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.087918 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.087962 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.087985 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-logs\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.088006 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.088068 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.088115 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.088154 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rztfj\" (UniqueName: \"kubernetes.io/projected/d2589e24-7bb6-49ac-a98c-55be590405df-kube-api-access-rztfj\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.088211 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.088273 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.088688 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.089329 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-logs\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.095942 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.098689 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.098896 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.107943 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.130297 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rztfj\" (UniqueName: \"kubernetes.io/projected/d2589e24-7bb6-49ac-a98c-55be590405df-kube-api-access-rztfj\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.190843 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.191233 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.201505 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.285074 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.376883 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.399221 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-config-data\") pod \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.399418 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-579mg\" (UniqueName: \"kubernetes.io/projected/e02b31ff-8f9a-4d65-b6fe-849630c89ede-kube-api-access-579mg\") pod \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.399456 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-credential-keys\") pod \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.399486 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-combined-ca-bundle\") pod \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.399616 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-scripts\") pod \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.399768 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-fernet-keys\") pod \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\" (UID: \"e02b31ff-8f9a-4d65-b6fe-849630c89ede\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.406788 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e02b31ff-8f9a-4d65-b6fe-849630c89ede" (UID: "e02b31ff-8f9a-4d65-b6fe-849630c89ede"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.409985 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e02b31ff-8f9a-4d65-b6fe-849630c89ede-kube-api-access-579mg" (OuterVolumeSpecName: "kube-api-access-579mg") pod "e02b31ff-8f9a-4d65-b6fe-849630c89ede" (UID: "e02b31ff-8f9a-4d65-b6fe-849630c89ede"). InnerVolumeSpecName "kube-api-access-579mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.413980 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e02b31ff-8f9a-4d65-b6fe-849630c89ede" (UID: "e02b31ff-8f9a-4d65-b6fe-849630c89ede"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.418004 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.436862 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-scripts" (OuterVolumeSpecName: "scripts") pod "e02b31ff-8f9a-4d65-b6fe-849630c89ede" (UID: "e02b31ff-8f9a-4d65-b6fe-849630c89ede"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.438409 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-config-data" (OuterVolumeSpecName: "config-data") pod "e02b31ff-8f9a-4d65-b6fe-849630c89ede" (UID: "e02b31ff-8f9a-4d65-b6fe-849630c89ede"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.484136 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.498841 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e02b31ff-8f9a-4d65-b6fe-849630c89ede" (UID: "e02b31ff-8f9a-4d65-b6fe-849630c89ede"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.501330 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f067613d-aad1-426c-858b-2753c83c725e\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.501505 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-combined-ca-bundle\") pod \"f067613d-aad1-426c-858b-2753c83c725e\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.501593 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-scripts\") pod \"f067613d-aad1-426c-858b-2753c83c725e\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.501744 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qdh8\" (UniqueName: \"kubernetes.io/projected/f067613d-aad1-426c-858b-2753c83c725e-kube-api-access-8qdh8\") pod \"f067613d-aad1-426c-858b-2753c83c725e\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.501877 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-logs\") pod \"f067613d-aad1-426c-858b-2753c83c725e\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.501995 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-config-data\") pod \"f067613d-aad1-426c-858b-2753c83c725e\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.502063 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-httpd-run\") pod \"f067613d-aad1-426c-858b-2753c83c725e\" (UID: \"f067613d-aad1-426c-858b-2753c83c725e\") " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.502566 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.502611 4675 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.502623 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.502634 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-579mg\" (UniqueName: \"kubernetes.io/projected/e02b31ff-8f9a-4d65-b6fe-849630c89ede-kube-api-access-579mg\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.502671 4675 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.502681 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02b31ff-8f9a-4d65-b6fe-849630c89ede-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.503359 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f067613d-aad1-426c-858b-2753c83c725e" (UID: "f067613d-aad1-426c-858b-2753c83c725e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.507032 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-logs" (OuterVolumeSpecName: "logs") pod "f067613d-aad1-426c-858b-2753c83c725e" (UID: "f067613d-aad1-426c-858b-2753c83c725e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.507132 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f067613d-aad1-426c-858b-2753c83c725e" (UID: "f067613d-aad1-426c-858b-2753c83c725e"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.513821 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f067613d-aad1-426c-858b-2753c83c725e-kube-api-access-8qdh8" (OuterVolumeSpecName: "kube-api-access-8qdh8") pod "f067613d-aad1-426c-858b-2753c83c725e" (UID: "f067613d-aad1-426c-858b-2753c83c725e"). InnerVolumeSpecName "kube-api-access-8qdh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.514877 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-scripts" (OuterVolumeSpecName: "scripts") pod "f067613d-aad1-426c-858b-2753c83c725e" (UID: "f067613d-aad1-426c-858b-2753c83c725e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.537785 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f067613d-aad1-426c-858b-2753c83c725e" (UID: "f067613d-aad1-426c-858b-2753c83c725e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.602809 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-config-data" (OuterVolumeSpecName: "config-data") pod "f067613d-aad1-426c-858b-2753c83c725e" (UID: "f067613d-aad1-426c-858b-2753c83c725e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.604295 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.604408 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.604478 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qdh8\" (UniqueName: \"kubernetes.io/projected/f067613d-aad1-426c-858b-2753c83c725e-kube-api-access-8qdh8\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.604540 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.604598 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f067613d-aad1-426c-858b-2753c83c725e-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.604665 4675 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f067613d-aad1-426c-858b-2753c83c725e-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.604737 4675 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.692311 4675 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.707715 4675 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.715012 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c47wh" event={"ID":"e02b31ff-8f9a-4d65-b6fe-849630c89ede","Type":"ContainerDied","Data":"32bb29adc6cdac74a5542fd993141d0764de6128ab4b1c2b4821edb2095216a3"} Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.715056 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32bb29adc6cdac74a5542fd993141d0764de6128ab4b1c2b4821edb2095216a3" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.715152 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c47wh" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.742229 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.744889 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f067613d-aad1-426c-858b-2753c83c725e","Type":"ContainerDied","Data":"c5911a6655248dfe10c3e155c0d7e7d40a3a8cf482fcbc79a4edf5299edb85fd"} Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.744944 4675 scope.go:117] "RemoveContainer" containerID="4c3b476dae250f0b670a91f41a0b331b91a5660d3ca825eee266f1114ed77b48" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.852822 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8594567c89-8xr7n"] Feb 19 09:02:00 crc kubenswrapper[4675]: E0219 09:02:00.853306 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e02b31ff-8f9a-4d65-b6fe-849630c89ede" containerName="keystone-bootstrap" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.853320 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="e02b31ff-8f9a-4d65-b6fe-849630c89ede" containerName="keystone-bootstrap" Feb 19 09:02:00 crc kubenswrapper[4675]: E0219 09:02:00.853332 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-httpd" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.853338 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-httpd" Feb 19 09:02:00 crc kubenswrapper[4675]: E0219 09:02:00.853369 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-log" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.853375 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-log" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.853576 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-log" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.853585 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="e02b31ff-8f9a-4d65-b6fe-849630c89ede" containerName="keystone-bootstrap" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.853596 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f067613d-aad1-426c-858b-2753c83c725e" containerName="glance-httpd" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.854237 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.864045 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.864458 4675 scope.go:117] "RemoveContainer" containerID="70230fe2d722bcf5790743c9b425ad6b79dbb41bea979ba0694852aecefabcfe" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.864519 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.864650 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.865177 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.865229 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6kbgs" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.865299 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.879692 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.898489 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.907691 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8594567c89-8xr7n"] Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.916730 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.918167 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.922690 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.923372 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 19 09:02:00 crc kubenswrapper[4675]: I0219 09:02:00.941091 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027279 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js5kg\" (UniqueName: \"kubernetes.io/projected/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-kube-api-access-js5kg\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027368 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-fernet-keys\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027398 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-internal-tls-certs\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027456 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027479 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027522 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027553 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027589 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-config-data\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027608 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-credential-keys\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027628 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-scripts\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027673 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-combined-ca-bundle\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027694 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-public-tls-certs\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027715 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027745 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlfsl\" (UniqueName: \"kubernetes.io/projected/c15d3213-8b81-4da3-851b-a6927b51f7aa-kube-api-access-xlfsl\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027779 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.027832 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.126375 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea0661b-0e96-4b12-af14-d5aab3d5edf9" path="/var/lib/kubelet/pods/6ea0661b-0e96-4b12-af14-d5aab3d5edf9/volumes" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.127545 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f067613d-aad1-426c-858b-2753c83c725e" path="/var/lib/kubelet/pods/f067613d-aad1-426c-858b-2753c83c725e/volumes" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.129417 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.129473 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js5kg\" (UniqueName: \"kubernetes.io/projected/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-kube-api-access-js5kg\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.129527 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-fernet-keys\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.129556 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-internal-tls-certs\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.129586 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.129609 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.130571 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.130657 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.130686 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-config-data\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.130712 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-credential-keys\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.130737 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-scripts\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.130935 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.132109 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-combined-ca-bundle\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.132152 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-public-tls-certs\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.132176 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.132193 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlfsl\" (UniqueName: \"kubernetes.io/projected/c15d3213-8b81-4da3-851b-a6927b51f7aa-kube-api-access-xlfsl\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.132238 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.136036 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.140310 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-internal-tls-certs\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.141059 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.141903 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.167675 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-fernet-keys\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.176276 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-credential-keys\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.177010 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.178074 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.183130 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-scripts\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.186904 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-public-tls-certs\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.187451 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-config-data\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.187804 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-combined-ca-bundle\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.189254 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlfsl\" (UniqueName: \"kubernetes.io/projected/c15d3213-8b81-4da3-851b-a6927b51f7aa-kube-api-access-xlfsl\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.189958 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.199231 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js5kg\" (UniqueName: \"kubernetes.io/projected/a6ae7ce1-9da5-49f4-8e32-0f1e0996080a-kube-api-access-js5kg\") pod \"keystone-8594567c89-8xr7n\" (UID: \"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a\") " pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.223438 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.225116 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:01 crc kubenswrapper[4675]: W0219 09:02:01.241756 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2589e24_7bb6_49ac_a98c_55be590405df.slice/crio-eb01db82e678b1693be4bb1c8fafc358cb1aa1729e5fc4bd23c784f1b61c6bdb WatchSource:0}: Error finding container eb01db82e678b1693be4bb1c8fafc358cb1aa1729e5fc4bd23c784f1b61c6bdb: Status 404 returned error can't find the container with id eb01db82e678b1693be4bb1c8fafc358cb1aa1729e5fc4bd23c784f1b61c6bdb Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.256321 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.378319 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wz69f" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.440251 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66sgh\" (UniqueName: \"kubernetes.io/projected/824b7d51-57cb-459f-8876-9d88025ff60f-kube-api-access-66sgh\") pod \"824b7d51-57cb-459f-8876-9d88025ff60f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.440364 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-scripts\") pod \"824b7d51-57cb-459f-8876-9d88025ff60f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.440613 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824b7d51-57cb-459f-8876-9d88025ff60f-logs\") pod \"824b7d51-57cb-459f-8876-9d88025ff60f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.440724 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-config-data\") pod \"824b7d51-57cb-459f-8876-9d88025ff60f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.442995 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824b7d51-57cb-459f-8876-9d88025ff60f-logs" (OuterVolumeSpecName: "logs") pod "824b7d51-57cb-459f-8876-9d88025ff60f" (UID: "824b7d51-57cb-459f-8876-9d88025ff60f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.445011 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-combined-ca-bundle\") pod \"824b7d51-57cb-459f-8876-9d88025ff60f\" (UID: \"824b7d51-57cb-459f-8876-9d88025ff60f\") " Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.445872 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/824b7d51-57cb-459f-8876-9d88025ff60f-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.468930 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-scripts" (OuterVolumeSpecName: "scripts") pod "824b7d51-57cb-459f-8876-9d88025ff60f" (UID: "824b7d51-57cb-459f-8876-9d88025ff60f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.469116 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824b7d51-57cb-459f-8876-9d88025ff60f-kube-api-access-66sgh" (OuterVolumeSpecName: "kube-api-access-66sgh") pod "824b7d51-57cb-459f-8876-9d88025ff60f" (UID: "824b7d51-57cb-459f-8876-9d88025ff60f"). InnerVolumeSpecName "kube-api-access-66sgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.490785 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-config-data" (OuterVolumeSpecName: "config-data") pod "824b7d51-57cb-459f-8876-9d88025ff60f" (UID: "824b7d51-57cb-459f-8876-9d88025ff60f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.502461 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "824b7d51-57cb-459f-8876-9d88025ff60f" (UID: "824b7d51-57cb-459f-8876-9d88025ff60f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.548469 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.548520 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.548533 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66sgh\" (UniqueName: \"kubernetes.io/projected/824b7d51-57cb-459f-8876-9d88025ff60f-kube-api-access-66sgh\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.548544 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824b7d51-57cb-459f-8876-9d88025ff60f-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.563977 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.781470 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wz69f" event={"ID":"824b7d51-57cb-459f-8876-9d88025ff60f","Type":"ContainerDied","Data":"976645a53bb49fa4cc8175fc33f99f623e36ab02248875abc6a2e21f7933230f"} Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.781897 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="976645a53bb49fa4cc8175fc33f99f623e36ab02248875abc6a2e21f7933230f" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.782046 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wz69f" Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.795500 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2589e24-7bb6-49ac-a98c-55be590405df","Type":"ContainerStarted","Data":"eb01db82e678b1693be4bb1c8fafc358cb1aa1729e5fc4bd23c784f1b61c6bdb"} Feb 19 09:02:01 crc kubenswrapper[4675]: I0219 09:02:01.894729 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8594567c89-8xr7n"] Feb 19 09:02:01 crc kubenswrapper[4675]: W0219 09:02:01.926345 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6ae7ce1_9da5_49f4_8e32_0f1e0996080a.slice/crio-93bf3fd5037736d1eb9c37e969831011047bff4abd0b93aedaa9d4b7f7564fbf WatchSource:0}: Error finding container 93bf3fd5037736d1eb9c37e969831011047bff4abd0b93aedaa9d4b7f7564fbf: Status 404 returned error can't find the container with id 93bf3fd5037736d1eb9c37e969831011047bff4abd0b93aedaa9d4b7f7564fbf Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.202236 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.741771 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7c56db696b-nwkrv"] Feb 19 09:02:02 crc kubenswrapper[4675]: E0219 09:02:02.742251 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824b7d51-57cb-459f-8876-9d88025ff60f" containerName="placement-db-sync" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.742272 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="824b7d51-57cb-459f-8876-9d88025ff60f" containerName="placement-db-sync" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.742626 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="824b7d51-57cb-459f-8876-9d88025ff60f" containerName="placement-db-sync" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.743961 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.746049 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d8w68" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.751598 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.751799 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.751915 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.752023 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.766023 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7c56db696b-nwkrv"] Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.828801 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.840411 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2589e24-7bb6-49ac-a98c-55be590405df","Type":"ContainerStarted","Data":"ec9226b305f0ce2cd0d2f932df6518705ee9d7c236b0bd4e5ad19b7f29b30918"} Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.843852 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c15d3213-8b81-4da3-851b-a6927b51f7aa","Type":"ContainerStarted","Data":"bc148688259049807e14c163f7d8fbc3891c7e1859418ea8ff56bb47a0c57596"} Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.858601 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8594567c89-8xr7n" event={"ID":"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a","Type":"ContainerStarted","Data":"ec2b987762109a3b41d157eb812c3e2111ef2bc41ed9b43d82609f52f773c629"} Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.858679 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8594567c89-8xr7n" event={"ID":"a6ae7ce1-9da5-49f4-8e32-0f1e0996080a","Type":"ContainerStarted","Data":"93bf3fd5037736d1eb9c37e969831011047bff4abd0b93aedaa9d4b7f7564fbf"} Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.859487 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.904509 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-combined-ca-bundle\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.904961 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-internal-tls-certs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.905063 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271c0509-6c25-44bf-8565-5d25488f67c7-logs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.905140 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-public-tls-certs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.907778 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-scripts\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.907852 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-config-data\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.907897 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fwl4\" (UniqueName: \"kubernetes.io/projected/271c0509-6c25-44bf-8565-5d25488f67c7-kube-api-access-8fwl4\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.940767 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-pj9fm"] Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.941137 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerName="dnsmasq-dns" containerID="cri-o://b48431ced98cfc828d5911429598b97b94ca60c2fb1c0a53790cdf84efa6180b" gracePeriod=10 Feb 19 09:02:02 crc kubenswrapper[4675]: I0219 09:02:02.953711 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8594567c89-8xr7n" podStartSLOduration=2.9536832840000002 podStartE2EDuration="2.953683284s" podCreationTimestamp="2026-02-19 09:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:02.895837192 +0000 UTC m=+1104.522927480" watchObservedRunningTime="2026-02-19 09:02:02.953683284 +0000 UTC m=+1104.580773552" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.010505 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-public-tls-certs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.010562 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-scripts\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.010595 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-config-data\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.010651 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fwl4\" (UniqueName: \"kubernetes.io/projected/271c0509-6c25-44bf-8565-5d25488f67c7-kube-api-access-8fwl4\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.015134 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-combined-ca-bundle\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.015287 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-internal-tls-certs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.015389 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271c0509-6c25-44bf-8565-5d25488f67c7-logs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.016156 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271c0509-6c25-44bf-8565-5d25488f67c7-logs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.023005 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-internal-tls-certs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.023728 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-scripts\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.025201 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-public-tls-certs\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.030100 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-config-data\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.032218 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fwl4\" (UniqueName: \"kubernetes.io/projected/271c0509-6c25-44bf-8565-5d25488f67c7-kube-api-access-8fwl4\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.043200 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271c0509-6c25-44bf-8565-5d25488f67c7-combined-ca-bundle\") pod \"placement-7c56db696b-nwkrv\" (UID: \"271c0509-6c25-44bf-8565-5d25488f67c7\") " pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.090152 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.879994 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c15d3213-8b81-4da3-851b-a6927b51f7aa","Type":"ContainerStarted","Data":"c966f779a7de37b0c7a32b6548053926814c76c76852307d77a6619f202d603a"} Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.884178 4675 generic.go:334] "Generic (PLEG): container finished" podID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerID="b48431ced98cfc828d5911429598b97b94ca60c2fb1c0a53790cdf84efa6180b" exitCode=0 Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.884234 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" event={"ID":"e69df799-9ba4-4336-9f8f-dc08facd7c1b","Type":"ContainerDied","Data":"b48431ced98cfc828d5911429598b97b94ca60c2fb1c0a53790cdf84efa6180b"} Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.887747 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2589e24-7bb6-49ac-a98c-55be590405df","Type":"ContainerStarted","Data":"c9173343cbea7deb828160fa7599a578a21600e26e0722f6d0dfaca422335b38"} Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.891609 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7c56db696b-nwkrv"] Feb 19 09:02:03 crc kubenswrapper[4675]: I0219 09:02:03.924655 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.924613101 podStartE2EDuration="4.924613101s" podCreationTimestamp="2026-02-19 09:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:03.913000859 +0000 UTC m=+1105.540091127" watchObservedRunningTime="2026-02-19 09:02:03.924613101 +0000 UTC m=+1105.551703389" Feb 19 09:02:04 crc kubenswrapper[4675]: E0219 09:02:04.564089 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0824f353_528c_4847_acfd_8f8635d6102f.slice/crio-0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d\": RecentStats: unable to find data in memory cache]" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.526357 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.681107 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-svc\") pod \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.681153 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-config\") pod \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.681288 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-swift-storage-0\") pod \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.681322 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7m8c\" (UniqueName: \"kubernetes.io/projected/e69df799-9ba4-4336-9f8f-dc08facd7c1b-kube-api-access-c7m8c\") pod \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.681370 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-nb\") pod \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.681390 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-sb\") pod \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\" (UID: \"e69df799-9ba4-4336-9f8f-dc08facd7c1b\") " Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.689878 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e69df799-9ba4-4336-9f8f-dc08facd7c1b-kube-api-access-c7m8c" (OuterVolumeSpecName: "kube-api-access-c7m8c") pod "e69df799-9ba4-4336-9f8f-dc08facd7c1b" (UID: "e69df799-9ba4-4336-9f8f-dc08facd7c1b"). InnerVolumeSpecName "kube-api-access-c7m8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.747837 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e69df799-9ba4-4336-9f8f-dc08facd7c1b" (UID: "e69df799-9ba4-4336-9f8f-dc08facd7c1b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.766362 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-config" (OuterVolumeSpecName: "config") pod "e69df799-9ba4-4336-9f8f-dc08facd7c1b" (UID: "e69df799-9ba4-4336-9f8f-dc08facd7c1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.785101 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.785140 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.785152 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7m8c\" (UniqueName: \"kubernetes.io/projected/e69df799-9ba4-4336-9f8f-dc08facd7c1b-kube-api-access-c7m8c\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.879984 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e69df799-9ba4-4336-9f8f-dc08facd7c1b" (UID: "e69df799-9ba4-4336-9f8f-dc08facd7c1b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.883339 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e69df799-9ba4-4336-9f8f-dc08facd7c1b" (UID: "e69df799-9ba4-4336-9f8f-dc08facd7c1b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.886970 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.887002 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.959969 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e69df799-9ba4-4336-9f8f-dc08facd7c1b" (UID: "e69df799-9ba4-4336-9f8f-dc08facd7c1b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.982938 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" event={"ID":"e69df799-9ba4-4336-9f8f-dc08facd7c1b","Type":"ContainerDied","Data":"b48fca0047cfbc68f0870aee9afb17e650fc2be1eab34ef5e4eade4be73c1e63"} Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.983028 4675 scope.go:117] "RemoveContainer" containerID="b48431ced98cfc828d5911429598b97b94ca60c2fb1c0a53790cdf84efa6180b" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.983974 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.988910 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e69df799-9ba4-4336-9f8f-dc08facd7c1b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:08 crc kubenswrapper[4675]: I0219 09:02:08.991579 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7c56db696b-nwkrv" event={"ID":"271c0509-6c25-44bf-8565-5d25488f67c7","Type":"ContainerStarted","Data":"7bae7bc09d33963debfe418629571fc2624939d80b3ca4526f7e7cdd61fcd79b"} Feb 19 09:02:09 crc kubenswrapper[4675]: I0219 09:02:09.029382 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-pj9fm"] Feb 19 09:02:09 crc kubenswrapper[4675]: I0219 09:02:09.041353 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-pj9fm"] Feb 19 09:02:09 crc kubenswrapper[4675]: I0219 09:02:09.044593 4675 scope.go:117] "RemoveContainer" containerID="48e1016435065fb0f5fc5e4aa7f2bc7700afdb6c45b8e62f4a0601755a9d2ad6" Feb 19 09:02:09 crc kubenswrapper[4675]: I0219 09:02:09.134425 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" path="/var/lib/kubelet/pods/e69df799-9ba4-4336-9f8f-dc08facd7c1b/volumes" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.000706 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wzzsx" event={"ID":"d78a6299-1816-4500-9e03-a004f52262f8","Type":"ContainerStarted","Data":"7f7a03e416d35621ac5cca2cbccaa49a47d259749f7a9fb584360690688e6bb2"} Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.006503 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c15d3213-8b81-4da3-851b-a6927b51f7aa","Type":"ContainerStarted","Data":"4491d6009caa431167f30a62aef780ddc463bf8377c2eda805101714b027a37a"} Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.009974 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbncs" event={"ID":"da9f4b70-4c48-4658-b2ae-b2d91b58f23c","Type":"ContainerStarted","Data":"0213029a75f69fc51b42589a3d914c1d1fdc5aab2b75fed314ed84ac198bedaf"} Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.012055 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerStarted","Data":"48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5"} Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.013867 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7c56db696b-nwkrv" event={"ID":"271c0509-6c25-44bf-8565-5d25488f67c7","Type":"ContainerStarted","Data":"362f06a1cd9bfe05c4462a61f5f835e00f5228b0a058c8281b73ac44161a62cc"} Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.013929 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7c56db696b-nwkrv" event={"ID":"271c0509-6c25-44bf-8565-5d25488f67c7","Type":"ContainerStarted","Data":"f57b1d80057e1813258a448cb9b50f851bf8c902b8c7a5804ed2bce2092f7e44"} Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.014002 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.014122 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.035789 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-wzzsx" podStartSLOduration=3.83459057 podStartE2EDuration="51.03575922s" podCreationTimestamp="2026-02-19 09:01:19 +0000 UTC" firstStartedPulling="2026-02-19 09:01:21.388201793 +0000 UTC m=+1063.015292061" lastFinishedPulling="2026-02-19 09:02:08.589370443 +0000 UTC m=+1110.216460711" observedRunningTime="2026-02-19 09:02:10.025149395 +0000 UTC m=+1111.652239663" watchObservedRunningTime="2026-02-19 09:02:10.03575922 +0000 UTC m=+1111.662849508" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.049005 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-cbncs" podStartSLOduration=4.1794218579999995 podStartE2EDuration="51.048978024s" podCreationTimestamp="2026-02-19 09:01:19 +0000 UTC" firstStartedPulling="2026-02-19 09:01:21.679753563 +0000 UTC m=+1063.306843841" lastFinishedPulling="2026-02-19 09:02:08.549309739 +0000 UTC m=+1110.176400007" observedRunningTime="2026-02-19 09:02:10.045215813 +0000 UTC m=+1111.672306081" watchObservedRunningTime="2026-02-19 09:02:10.048978024 +0000 UTC m=+1111.676068292" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.059832 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.082673 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.082635217 podStartE2EDuration="10.082635217s" podCreationTimestamp="2026-02-19 09:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:10.07232411 +0000 UTC m=+1111.699414398" watchObservedRunningTime="2026-02-19 09:02:10.082635217 +0000 UTC m=+1111.709725485" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.100486 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7c56db696b-nwkrv" podStartSLOduration=8.100462755 podStartE2EDuration="8.100462755s" podCreationTimestamp="2026-02-19 09:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:10.089965593 +0000 UTC m=+1111.717055861" watchObservedRunningTime="2026-02-19 09:02:10.100462755 +0000 UTC m=+1111.727553043" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.192328 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64d55559f4-zwfgc" podUID="18d90d12-7724-438d-8cd4-9fbe70c8ee89" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.484555 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.484601 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.514482 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.526571 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 19 09:02:10 crc kubenswrapper[4675]: I0219 09:02:10.805025 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-fcfdd6f9f-pj9fm" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: i/o timeout" Feb 19 09:02:11 crc kubenswrapper[4675]: I0219 09:02:11.026547 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 19 09:02:11 crc kubenswrapper[4675]: I0219 09:02:11.026586 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 19 09:02:11 crc kubenswrapper[4675]: I0219 09:02:11.565130 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:11 crc kubenswrapper[4675]: I0219 09:02:11.565220 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:11 crc kubenswrapper[4675]: I0219 09:02:11.623509 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:11 crc kubenswrapper[4675]: I0219 09:02:11.663927 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:12 crc kubenswrapper[4675]: I0219 09:02:12.035098 4675 generic.go:334] "Generic (PLEG): container finished" podID="da9f4b70-4c48-4658-b2ae-b2d91b58f23c" containerID="0213029a75f69fc51b42589a3d914c1d1fdc5aab2b75fed314ed84ac198bedaf" exitCode=0 Feb 19 09:02:12 crc kubenswrapper[4675]: I0219 09:02:12.036265 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbncs" event={"ID":"da9f4b70-4c48-4658-b2ae-b2d91b58f23c","Type":"ContainerDied","Data":"0213029a75f69fc51b42589a3d914c1d1fdc5aab2b75fed314ed84ac198bedaf"} Feb 19 09:02:12 crc kubenswrapper[4675]: I0219 09:02:12.036772 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:12 crc kubenswrapper[4675]: I0219 09:02:12.036814 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.044888 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.044925 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.210845 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.458479 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.604872 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbncs" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.683323 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-combined-ca-bundle\") pod \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.683470 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-db-sync-config-data\") pod \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.683549 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z27dc\" (UniqueName: \"kubernetes.io/projected/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-kube-api-access-z27dc\") pod \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\" (UID: \"da9f4b70-4c48-4658-b2ae-b2d91b58f23c\") " Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.691695 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "da9f4b70-4c48-4658-b2ae-b2d91b58f23c" (UID: "da9f4b70-4c48-4658-b2ae-b2d91b58f23c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.691876 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-kube-api-access-z27dc" (OuterVolumeSpecName: "kube-api-access-z27dc") pod "da9f4b70-4c48-4658-b2ae-b2d91b58f23c" (UID: "da9f4b70-4c48-4658-b2ae-b2d91b58f23c"). InnerVolumeSpecName "kube-api-access-z27dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.736782 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da9f4b70-4c48-4658-b2ae-b2d91b58f23c" (UID: "da9f4b70-4c48-4658-b2ae-b2d91b58f23c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.786130 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z27dc\" (UniqueName: \"kubernetes.io/projected/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-kube-api-access-z27dc\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.786169 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:13 crc kubenswrapper[4675]: I0219 09:02:13.786180 4675 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/da9f4b70-4c48-4658-b2ae-b2d91b58f23c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.052848 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbncs" event={"ID":"da9f4b70-4c48-4658-b2ae-b2d91b58f23c","Type":"ContainerDied","Data":"6b6a8eb1f8e8ebdb184ed9ff853aa6c56caed54aba045c50eb0f8c7e50f35705"} Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.052895 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b6a8eb1f8e8ebdb184ed9ff853aa6c56caed54aba045c50eb0f8c7e50f35705" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.052911 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbncs" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.053006 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.609659 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.968536 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7667dd7745-49ztz"] Feb 19 09:02:14 crc kubenswrapper[4675]: E0219 09:02:14.969995 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerName="dnsmasq-dns" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.970201 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerName="dnsmasq-dns" Feb 19 09:02:14 crc kubenswrapper[4675]: E0219 09:02:14.970404 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da9f4b70-4c48-4658-b2ae-b2d91b58f23c" containerName="barbican-db-sync" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.970848 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="da9f4b70-4c48-4658-b2ae-b2d91b58f23c" containerName="barbican-db-sync" Feb 19 09:02:14 crc kubenswrapper[4675]: E0219 09:02:14.971293 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerName="init" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.976210 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerName="init" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.976824 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="e69df799-9ba4-4336-9f8f-dc08facd7c1b" containerName="dnsmasq-dns" Feb 19 09:02:14 crc kubenswrapper[4675]: I0219 09:02:14.976946 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="da9f4b70-4c48-4658-b2ae-b2d91b58f23c" containerName="barbican-db-sync" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.011014 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.027277 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.029615 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.029838 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mqn6h" Feb 19 09:02:15 crc kubenswrapper[4675]: E0219 09:02:15.037854 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0824f353_528c_4847_acfd_8f8635d6102f.slice/crio-0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d\": RecentStats: unable to find data in memory cache]" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.061166 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7667dd7745-49ztz"] Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.136276 4675 generic.go:334] "Generic (PLEG): container finished" podID="d78a6299-1816-4500-9e03-a004f52262f8" containerID="7f7a03e416d35621ac5cca2cbccaa49a47d259749f7a9fb584360690688e6bb2" exitCode=0 Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.139433 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-bcb894f54-l299f"] Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.145849 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wzzsx" event={"ID":"d78a6299-1816-4500-9e03-a004f52262f8","Type":"ContainerDied","Data":"7f7a03e416d35621ac5cca2cbccaa49a47d259749f7a9fb584360690688e6bb2"} Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.145920 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-bcb894f54-l299f"] Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.146064 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.146179 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-b2std"] Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.152242 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-config-data\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.152385 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-combined-ca-bundle\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.152437 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a4a2232-762f-4fc1-83fd-45d33b408e5b-logs\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.152562 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pr7x\" (UniqueName: \"kubernetes.io/projected/5a4a2232-762f-4fc1-83fd-45d33b408e5b-kube-api-access-2pr7x\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.152691 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-config-data-custom\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.153575 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.155276 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.169244 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-b2std"] Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.209026 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-77b588c78-htngd"] Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.211238 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.218343 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.249835 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77b588c78-htngd"] Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.254753 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-config-data\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.254816 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-config-data-custom\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.254845 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-combined-ca-bundle\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.254891 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-config-data\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.254937 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpzph\" (UniqueName: \"kubernetes.io/projected/8649bd53-643b-466d-a528-45d2aa799316-kube-api-access-jpzph\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.254962 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-combined-ca-bundle\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255199 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs9wb\" (UniqueName: \"kubernetes.io/projected/e4ff5bfa-5771-4076-8be7-ce62be77af12-kube-api-access-hs9wb\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255236 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-config-data-custom\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255264 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a4a2232-762f-4fc1-83fd-45d33b408e5b-logs\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255282 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8649bd53-643b-466d-a528-45d2aa799316-logs\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255307 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pr7x\" (UniqueName: \"kubernetes.io/projected/5a4a2232-762f-4fc1-83fd-45d33b408e5b-kube-api-access-2pr7x\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255325 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ff5bfa-5771-4076-8be7-ce62be77af12-logs\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255345 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data-custom\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255366 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-combined-ca-bundle\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.255397 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.256910 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a4a2232-762f-4fc1-83fd-45d33b408e5b-logs\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.268314 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-combined-ca-bundle\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.283109 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-config-data-custom\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.303902 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a4a2232-762f-4fc1-83fd-45d33b408e5b-config-data\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.312575 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pr7x\" (UniqueName: \"kubernetes.io/projected/5a4a2232-762f-4fc1-83fd-45d33b408e5b-kube-api-access-2pr7x\") pod \"barbican-worker-7667dd7745-49ztz\" (UID: \"5a4a2232-762f-4fc1-83fd-45d33b408e5b\") " pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358190 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs9wb\" (UniqueName: \"kubernetes.io/projected/e4ff5bfa-5771-4076-8be7-ce62be77af12-kube-api-access-hs9wb\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358249 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358270 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358298 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-config-data-custom\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358330 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8649bd53-643b-466d-a528-45d2aa799316-logs\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358352 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-svc\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358368 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n7nm\" (UniqueName: \"kubernetes.io/projected/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-kube-api-access-2n7nm\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358392 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ff5bfa-5771-4076-8be7-ce62be77af12-logs\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358416 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data-custom\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358437 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-config\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358465 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-combined-ca-bundle\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358491 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358514 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358581 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-config-data\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358634 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-combined-ca-bundle\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.358713 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpzph\" (UniqueName: \"kubernetes.io/projected/8649bd53-643b-466d-a528-45d2aa799316-kube-api-access-jpzph\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.360767 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8649bd53-643b-466d-a528-45d2aa799316-logs\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.360912 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ff5bfa-5771-4076-8be7-ce62be77af12-logs\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.364743 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-config-data\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.368530 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-combined-ca-bundle\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.369043 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-combined-ca-bundle\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.371304 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8649bd53-643b-466d-a528-45d2aa799316-config-data-custom\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.373244 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data-custom\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.379957 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.388115 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpzph\" (UniqueName: \"kubernetes.io/projected/8649bd53-643b-466d-a528-45d2aa799316-kube-api-access-jpzph\") pod \"barbican-keystone-listener-bcb894f54-l299f\" (UID: \"8649bd53-643b-466d-a528-45d2aa799316\") " pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.396740 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs9wb\" (UniqueName: \"kubernetes.io/projected/e4ff5bfa-5771-4076-8be7-ce62be77af12-kube-api-access-hs9wb\") pod \"barbican-api-77b588c78-htngd\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.406319 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7667dd7745-49ztz" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.459890 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.460270 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.460317 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-svc\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.460335 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n7nm\" (UniqueName: \"kubernetes.io/projected/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-kube-api-access-2n7nm\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.460366 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-config\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.460404 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.461036 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.461345 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-svc\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.461999 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.462328 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-config\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.462886 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.478031 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n7nm\" (UniqueName: \"kubernetes.io/projected/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-kube-api-access-2n7nm\") pod \"dnsmasq-dns-688c87cc99-b2std\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.493838 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-bcb894f54-l299f" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.515011 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:15 crc kubenswrapper[4675]: I0219 09:02:15.552110 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:16 crc kubenswrapper[4675]: I0219 09:02:16.780129 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.764499 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b4d586cf4-9nm5g"] Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.772722 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.782490 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.782809 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.845058 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b4d586cf4-9nm5g"] Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.925854 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-combined-ca-bundle\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.925918 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c045dff4-aad5-44b8-a942-20a45f927859-logs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.925940 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-config-data\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.926109 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-config-data-custom\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.926153 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-internal-tls-certs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.926346 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-public-tls-certs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:17 crc kubenswrapper[4675]: I0219 09:02:17.926560 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p28gt\" (UniqueName: \"kubernetes.io/projected/c045dff4-aad5-44b8-a942-20a45f927859-kube-api-access-p28gt\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.028794 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-internal-tls-certs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.028868 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-public-tls-certs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.028924 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p28gt\" (UniqueName: \"kubernetes.io/projected/c045dff4-aad5-44b8-a942-20a45f927859-kube-api-access-p28gt\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.028983 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-combined-ca-bundle\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.029006 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c045dff4-aad5-44b8-a942-20a45f927859-logs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.029024 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-config-data\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.029068 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-config-data-custom\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.030779 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c045dff4-aad5-44b8-a942-20a45f927859-logs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.039109 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-public-tls-certs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.044913 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p28gt\" (UniqueName: \"kubernetes.io/projected/c045dff4-aad5-44b8-a942-20a45f927859-kube-api-access-p28gt\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.049294 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-internal-tls-certs\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.054763 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-config-data-custom\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.064688 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-config-data\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.065234 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c045dff4-aad5-44b8-a942-20a45f927859-combined-ca-bundle\") pod \"barbican-api-6b4d586cf4-9nm5g\" (UID: \"c045dff4-aad5-44b8-a942-20a45f927859\") " pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:18 crc kubenswrapper[4675]: I0219 09:02:18.112707 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.127757 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.184838 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wzzsx" event={"ID":"d78a6299-1816-4500-9e03-a004f52262f8","Type":"ContainerDied","Data":"81a65ebb8419dc93179494ed3fb0e97297ea368e793396b2ea5f51ea3c4ad701"} Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.184891 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a65ebb8419dc93179494ed3fb0e97297ea368e793396b2ea5f51ea3c4ad701" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.184893 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wzzsx" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.190706 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64d55559f4-zwfgc" podUID="18d90d12-7724-438d-8cd4-9fbe70c8ee89" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.274847 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvdds\" (UniqueName: \"kubernetes.io/projected/d78a6299-1816-4500-9e03-a004f52262f8-kube-api-access-xvdds\") pod \"d78a6299-1816-4500-9e03-a004f52262f8\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.274930 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-db-sync-config-data\") pod \"d78a6299-1816-4500-9e03-a004f52262f8\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.275037 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-config-data\") pod \"d78a6299-1816-4500-9e03-a004f52262f8\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.275060 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-scripts\") pod \"d78a6299-1816-4500-9e03-a004f52262f8\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.275079 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-combined-ca-bundle\") pod \"d78a6299-1816-4500-9e03-a004f52262f8\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.275205 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d78a6299-1816-4500-9e03-a004f52262f8-etc-machine-id\") pod \"d78a6299-1816-4500-9e03-a004f52262f8\" (UID: \"d78a6299-1816-4500-9e03-a004f52262f8\") " Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.275364 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d78a6299-1816-4500-9e03-a004f52262f8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d78a6299-1816-4500-9e03-a004f52262f8" (UID: "d78a6299-1816-4500-9e03-a004f52262f8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.275685 4675 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d78a6299-1816-4500-9e03-a004f52262f8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.281490 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d78a6299-1816-4500-9e03-a004f52262f8" (UID: "d78a6299-1816-4500-9e03-a004f52262f8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.282234 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-scripts" (OuterVolumeSpecName: "scripts") pod "d78a6299-1816-4500-9e03-a004f52262f8" (UID: "d78a6299-1816-4500-9e03-a004f52262f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.294136 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d78a6299-1816-4500-9e03-a004f52262f8-kube-api-access-xvdds" (OuterVolumeSpecName: "kube-api-access-xvdds") pod "d78a6299-1816-4500-9e03-a004f52262f8" (UID: "d78a6299-1816-4500-9e03-a004f52262f8"). InnerVolumeSpecName "kube-api-access-xvdds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.312132 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d78a6299-1816-4500-9e03-a004f52262f8" (UID: "d78a6299-1816-4500-9e03-a004f52262f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.353706 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-config-data" (OuterVolumeSpecName: "config-data") pod "d78a6299-1816-4500-9e03-a004f52262f8" (UID: "d78a6299-1816-4500-9e03-a004f52262f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.377224 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvdds\" (UniqueName: \"kubernetes.io/projected/d78a6299-1816-4500-9e03-a004f52262f8-kube-api-access-xvdds\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.377257 4675 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.377267 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.377277 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:20 crc kubenswrapper[4675]: I0219 09:02:20.377286 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d78a6299-1816-4500-9e03-a004f52262f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.126597 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b4d586cf4-9nm5g"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.127087 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77b588c78-htngd"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.222277 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7667dd7745-49ztz"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.471604 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:21 crc kubenswrapper[4675]: E0219 09:02:21.472154 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d78a6299-1816-4500-9e03-a004f52262f8" containerName="cinder-db-sync" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.472173 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d78a6299-1816-4500-9e03-a004f52262f8" containerName="cinder-db-sync" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.472346 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d78a6299-1816-4500-9e03-a004f52262f8" containerName="cinder-db-sync" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.473292 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: W0219 09:02:21.475605 4675 reflector.go:561] object-"openstack"/"cinder-scheduler-config-data": failed to list *v1.Secret: secrets "cinder-scheduler-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Feb 19 09:02:21 crc kubenswrapper[4675]: E0219 09:02:21.475687 4675 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cinder-scheduler-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cinder-scheduler-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.475783 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.475952 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fhphd" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.476085 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.552039 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.589918 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-b2std"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.601973 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.602026 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.602048 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.602103 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-scripts\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.602143 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fzzn\" (UniqueName: \"kubernetes.io/projected/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-kube-api-access-7fzzn\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.602166 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.641906 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qr6ht"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.644059 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.696058 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qr6ht"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.703661 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704140 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-scripts\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704255 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fzzn\" (UniqueName: \"kubernetes.io/projected/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-kube-api-access-7fzzn\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704343 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704546 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704702 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44v27\" (UniqueName: \"kubernetes.io/projected/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-kube-api-access-44v27\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704828 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704899 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.704976 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-config\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.705056 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.705221 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.705767 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.707753 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.713982 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.723587 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fzzn\" (UniqueName: \"kubernetes.io/projected/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-kube-api-access-7fzzn\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.737556 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.738239 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-scripts\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.762680 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.764323 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.766880 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.774097 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:21 crc kubenswrapper[4675]: W0219 09:02:21.795495 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc045dff4_aad5_44b8_a942_20a45f927859.slice/crio-190fe3b101c3868730d1ce51d6c30731d7427a3efbb851170fb8ee0c2ea13a98 WatchSource:0}: Error finding container 190fe3b101c3868730d1ce51d6c30731d7427a3efbb851170fb8ee0c2ea13a98: Status 404 returned error can't find the container with id 190fe3b101c3868730d1ce51d6c30731d7427a3efbb851170fb8ee0c2ea13a98 Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.807120 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.807185 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44v27\" (UniqueName: \"kubernetes.io/projected/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-kube-api-access-44v27\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.807245 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.807264 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.807287 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-config\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.807343 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.808930 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.809040 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-config\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.809301 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.809673 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.810172 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.833160 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44v27\" (UniqueName: \"kubernetes.io/projected/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-kube-api-access-44v27\") pod \"dnsmasq-dns-6bb4fc677f-qr6ht\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.910371 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.914710 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwsvk\" (UniqueName: \"kubernetes.io/projected/457624d6-3a3d-406f-bc5e-626f1a0244ac-kube-api-access-nwsvk\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.915007 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data-custom\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.915288 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/457624d6-3a3d-406f-bc5e-626f1a0244ac-logs\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.915631 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.915805 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/457624d6-3a3d-406f-bc5e-626f1a0244ac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.915979 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-scripts\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:21 crc kubenswrapper[4675]: I0219 09:02:21.983163 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017413 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/457624d6-3a3d-406f-bc5e-626f1a0244ac-logs\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017456 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017487 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/457624d6-3a3d-406f-bc5e-626f1a0244ac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017535 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-scripts\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017625 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017682 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwsvk\" (UniqueName: \"kubernetes.io/projected/457624d6-3a3d-406f-bc5e-626f1a0244ac-kube-api-access-nwsvk\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017700 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data-custom\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.017879 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/457624d6-3a3d-406f-bc5e-626f1a0244ac-logs\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.022525 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/457624d6-3a3d-406f-bc5e-626f1a0244ac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.023490 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data-custom\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.026963 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-scripts\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.038235 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.044854 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.065035 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwsvk\" (UniqueName: \"kubernetes.io/projected/457624d6-3a3d-406f-bc5e-626f1a0244ac-kube-api-access-nwsvk\") pod \"cinder-api-0\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.183816 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-bcb894f54-l299f"] Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.224524 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.235783 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77b588c78-htngd" event={"ID":"e4ff5bfa-5771-4076-8be7-ce62be77af12","Type":"ContainerStarted","Data":"766445bb13fa4bb76aa03bc764fe6e0ed98768419a7a601565e2b59dac045d04"} Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.236081 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77b588c78-htngd" event={"ID":"e4ff5bfa-5771-4076-8be7-ce62be77af12","Type":"ContainerStarted","Data":"fde35d31b251aadfcc817ebe17c59402316931657d2603038661d4b66a879a03"} Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.238148 4675 generic.go:334] "Generic (PLEG): container finished" podID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerID="80ec74317bd11d88956bcf2d002aa1462a757c1812cf26956b46d048b2b76569" exitCode=137 Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.238167 4675 generic.go:334] "Generic (PLEG): container finished" podID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerID="6c1fd05e0d490ce48f6728f5ab42948031b2d6da4ff8ab7aba93446c7a42a968" exitCode=137 Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.238197 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bf577555-t6p57" event={"ID":"d2e38966-737c-49a6-800a-3293f1fd65a9","Type":"ContainerDied","Data":"80ec74317bd11d88956bcf2d002aa1462a757c1812cf26956b46d048b2b76569"} Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.238216 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bf577555-t6p57" event={"ID":"d2e38966-737c-49a6-800a-3293f1fd65a9","Type":"ContainerDied","Data":"6c1fd05e0d490ce48f6728f5ab42948031b2d6da4ff8ab7aba93446c7a42a968"} Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.239277 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7667dd7745-49ztz" event={"ID":"5a4a2232-762f-4fc1-83fd-45d33b408e5b","Type":"ContainerStarted","Data":"7d65b2dfd535103a7e54672dc5b828312a380a8f987ecf8f431ecaf1a458d839"} Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.241199 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b4d586cf4-9nm5g" event={"ID":"c045dff4-aad5-44b8-a942-20a45f927859","Type":"ContainerStarted","Data":"42238b43608c281e823e6dddb3e3da864a619d99289694a89901a8b66600824a"} Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.241221 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b4d586cf4-9nm5g" event={"ID":"c045dff4-aad5-44b8-a942-20a45f927859","Type":"ContainerStarted","Data":"190fe3b101c3868730d1ce51d6c30731d7427a3efbb851170fb8ee0c2ea13a98"} Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.294539 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-b2std"] Feb 19 09:02:22 crc kubenswrapper[4675]: W0219 09:02:22.467606 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b94f207_4b1c_49ff_a2c2_ff15d94557a6.slice/crio-e789b52065f282758d82e4a654ebec0192dd27b87a2075fec25ca429a6903571 WatchSource:0}: Error finding container e789b52065f282758d82e4a654ebec0192dd27b87a2075fec25ca429a6903571: Status 404 returned error can't find the container with id e789b52065f282758d82e4a654ebec0192dd27b87a2075fec25ca429a6903571 Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.595756 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.602794 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.681336 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qr6ht"] Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.704238 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.737542 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.932525 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.956216 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:02:22 crc kubenswrapper[4675]: I0219 09:02:22.992317 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044322 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2e38966-737c-49a6-800a-3293f1fd65a9-horizon-secret-key\") pod \"d2e38966-737c-49a6-800a-3293f1fd65a9\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044374 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-scripts\") pod \"d2e38966-737c-49a6-800a-3293f1fd65a9\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044439 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f421816-e18a-4955-ba7a-557ed5e91ad3-logs\") pod \"0f421816-e18a-4955-ba7a-557ed5e91ad3\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044526 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqjhk\" (UniqueName: \"kubernetes.io/projected/d2e38966-737c-49a6-800a-3293f1fd65a9-kube-api-access-kqjhk\") pod \"d2e38966-737c-49a6-800a-3293f1fd65a9\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044554 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-scripts\") pod \"0f421816-e18a-4955-ba7a-557ed5e91ad3\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044571 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f421816-e18a-4955-ba7a-557ed5e91ad3-horizon-secret-key\") pod \"0f421816-e18a-4955-ba7a-557ed5e91ad3\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044633 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-config-data\") pod \"0f421816-e18a-4955-ba7a-557ed5e91ad3\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044710 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk9d7\" (UniqueName: \"kubernetes.io/projected/0f421816-e18a-4955-ba7a-557ed5e91ad3-kube-api-access-lk9d7\") pod \"0f421816-e18a-4955-ba7a-557ed5e91ad3\" (UID: \"0f421816-e18a-4955-ba7a-557ed5e91ad3\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044735 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-config-data\") pod \"d2e38966-737c-49a6-800a-3293f1fd65a9\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.044764 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2e38966-737c-49a6-800a-3293f1fd65a9-logs\") pod \"d2e38966-737c-49a6-800a-3293f1fd65a9\" (UID: \"d2e38966-737c-49a6-800a-3293f1fd65a9\") " Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.045781 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2e38966-737c-49a6-800a-3293f1fd65a9-logs" (OuterVolumeSpecName: "logs") pod "d2e38966-737c-49a6-800a-3293f1fd65a9" (UID: "d2e38966-737c-49a6-800a-3293f1fd65a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.060743 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f421816-e18a-4955-ba7a-557ed5e91ad3-logs" (OuterVolumeSpecName: "logs") pod "0f421816-e18a-4955-ba7a-557ed5e91ad3" (UID: "0f421816-e18a-4955-ba7a-557ed5e91ad3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.082287 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e38966-737c-49a6-800a-3293f1fd65a9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d2e38966-737c-49a6-800a-3293f1fd65a9" (UID: "d2e38966-737c-49a6-800a-3293f1fd65a9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.082479 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f421816-e18a-4955-ba7a-557ed5e91ad3-kube-api-access-lk9d7" (OuterVolumeSpecName: "kube-api-access-lk9d7") pod "0f421816-e18a-4955-ba7a-557ed5e91ad3" (UID: "0f421816-e18a-4955-ba7a-557ed5e91ad3"). InnerVolumeSpecName "kube-api-access-lk9d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.085316 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f421816-e18a-4955-ba7a-557ed5e91ad3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0f421816-e18a-4955-ba7a-557ed5e91ad3" (UID: "0f421816-e18a-4955-ba7a-557ed5e91ad3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.085500 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e38966-737c-49a6-800a-3293f1fd65a9-kube-api-access-kqjhk" (OuterVolumeSpecName: "kube-api-access-kqjhk") pod "d2e38966-737c-49a6-800a-3293f1fd65a9" (UID: "d2e38966-737c-49a6-800a-3293f1fd65a9"). InnerVolumeSpecName "kube-api-access-kqjhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.098137 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.119178 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-config-data" (OuterVolumeSpecName: "config-data") pod "d2e38966-737c-49a6-800a-3293f1fd65a9" (UID: "d2e38966-737c-49a6-800a-3293f1fd65a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.134432 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-scripts" (OuterVolumeSpecName: "scripts") pod "d2e38966-737c-49a6-800a-3293f1fd65a9" (UID: "d2e38966-737c-49a6-800a-3293f1fd65a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.134839 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-config-data" (OuterVolumeSpecName: "config-data") pod "0f421816-e18a-4955-ba7a-557ed5e91ad3" (UID: "0f421816-e18a-4955-ba7a-557ed5e91ad3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.138043 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-scripts" (OuterVolumeSpecName: "scripts") pod "0f421816-e18a-4955-ba7a-557ed5e91ad3" (UID: "0f421816-e18a-4955-ba7a-557ed5e91ad3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150001 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk9d7\" (UniqueName: \"kubernetes.io/projected/0f421816-e18a-4955-ba7a-557ed5e91ad3-kube-api-access-lk9d7\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150031 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150042 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2e38966-737c-49a6-800a-3293f1fd65a9-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150055 4675 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2e38966-737c-49a6-800a-3293f1fd65a9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150063 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2e38966-737c-49a6-800a-3293f1fd65a9-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150071 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f421816-e18a-4955-ba7a-557ed5e91ad3-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150089 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqjhk\" (UniqueName: \"kubernetes.io/projected/d2e38966-737c-49a6-800a-3293f1fd65a9-kube-api-access-kqjhk\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150096 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150104 4675 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0f421816-e18a-4955-ba7a-557ed5e91ad3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.150113 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f421816-e18a-4955-ba7a-557ed5e91ad3-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.267319 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-bcb894f54-l299f" event={"ID":"8649bd53-643b-466d-a528-45d2aa799316","Type":"ContainerStarted","Data":"69f97c4e276c295edd065eca2fbaed20e415987eb63c4c12d7ef1cb4ecf35fb3"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.270772 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77b588c78-htngd" event={"ID":"e4ff5bfa-5771-4076-8be7-ce62be77af12","Type":"ContainerStarted","Data":"113cf947778153c5ef2210ccd374f45ccfefe3b42bbb558eb51f3361c9d4dc43"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.271820 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.271893 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.282570 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerStarted","Data":"77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.282809 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-central-agent" containerID="cri-o://6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac" gracePeriod=30 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.283154 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.283222 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="proxy-httpd" containerID="cri-o://77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2" gracePeriod=30 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.283290 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="sg-core" containerID="cri-o://48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5" gracePeriod=30 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.283402 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-notification-agent" containerID="cri-o://3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e" gracePeriod=30 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.290764 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-77b588c78-htngd" podStartSLOduration=8.290740781 podStartE2EDuration="8.290740781s" podCreationTimestamp="2026-02-19 09:02:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:23.28845867 +0000 UTC m=+1124.915548948" watchObservedRunningTime="2026-02-19 09:02:23.290740781 +0000 UTC m=+1124.917831049" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.300536 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b4d586cf4-9nm5g" event={"ID":"c045dff4-aad5-44b8-a942-20a45f927859","Type":"ContainerStarted","Data":"2bf43bc633f03343d9b95736901b5d3abb3f9c123bc37ff55ddc1f05299b4788"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.301579 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.301616 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.309708 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"457624d6-3a3d-406f-bc5e-626f1a0244ac","Type":"ContainerStarted","Data":"b79b16cede22a55cbaa6ee0b64f899325cf976149e485661d3144dd0a755d270"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.316011 4675 generic.go:334] "Generic (PLEG): container finished" podID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerID="006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad" exitCode=137 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.316048 4675 generic.go:334] "Generic (PLEG): container finished" podID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerID="40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5" exitCode=137 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.316094 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dfb48549c-lkjnx" event={"ID":"0f421816-e18a-4955-ba7a-557ed5e91ad3","Type":"ContainerDied","Data":"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.316125 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dfb48549c-lkjnx" event={"ID":"0f421816-e18a-4955-ba7a-557ed5e91ad3","Type":"ContainerDied","Data":"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.316136 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dfb48549c-lkjnx" event={"ID":"0f421816-e18a-4955-ba7a-557ed5e91ad3","Type":"ContainerDied","Data":"cb90b81a89ea792092d280f5c90c45369f10b6cfb73593e37fb24a9eecfb5241"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.316151 4675 scope.go:117] "RemoveContainer" containerID="006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.316291 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dfb48549c-lkjnx" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.324543 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" event={"ID":"825f8331-17ce-4e8e-beb5-4f2807fc4ceb","Type":"ContainerStarted","Data":"f21a39d1faebc8547eb29ffb30a6ca416e32d59c7493d37d795c1a8a3f16a1c0"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.347128 4675 generic.go:334] "Generic (PLEG): container finished" podID="1b94f207-4b1c-49ff-a2c2-ff15d94557a6" containerID="dd2513ddf92af792268c2797c706db1f07d9fc7ea2032eb00819672fe96c5b48" exitCode=0 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.347276 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-b2std" event={"ID":"1b94f207-4b1c-49ff-a2c2-ff15d94557a6","Type":"ContainerDied","Data":"dd2513ddf92af792268c2797c706db1f07d9fc7ea2032eb00819672fe96c5b48"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.347318 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-b2std" event={"ID":"1b94f207-4b1c-49ff-a2c2-ff15d94557a6","Type":"ContainerStarted","Data":"e789b52065f282758d82e4a654ebec0192dd27b87a2075fec25ca429a6903571"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.350502 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.98647018 podStartE2EDuration="1m4.350475013s" podCreationTimestamp="2026-02-19 09:01:19 +0000 UTC" firstStartedPulling="2026-02-19 09:01:21.518819324 +0000 UTC m=+1063.145909592" lastFinishedPulling="2026-02-19 09:02:21.882824157 +0000 UTC m=+1123.509914425" observedRunningTime="2026-02-19 09:02:23.324158008 +0000 UTC m=+1124.951248276" watchObservedRunningTime="2026-02-19 09:02:23.350475013 +0000 UTC m=+1124.977565281" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.377367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bf577555-t6p57" event={"ID":"d2e38966-737c-49a6-800a-3293f1fd65a9","Type":"ContainerDied","Data":"991b0cde90402a90d71ae6b88a533a27c71375fb74cad19cb493296e5486801e"} Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.377478 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bf577555-t6p57" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.387367 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b4d586cf4-9nm5g" podStartSLOduration=6.387311471 podStartE2EDuration="6.387311471s" podCreationTimestamp="2026-02-19 09:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:23.378184706 +0000 UTC m=+1125.005274974" watchObservedRunningTime="2026-02-19 09:02:23.387311471 +0000 UTC m=+1125.014401739" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.480565 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.511824 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6dfb48549c-lkjnx"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.527676 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6dfb48549c-lkjnx"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.594768 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65bf577555-t6p57"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.607034 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-65bf577555-t6p57"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.624362 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7db6784559-bx2jh"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.624618 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7db6784559-bx2jh" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-api" containerID="cri-o://bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28" gracePeriod=30 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.625129 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7db6784559-bx2jh" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-httpd" containerID="cri-o://aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02" gracePeriod=30 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635150 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-57f4c8679f-grtkm"] Feb 19 09:02:23 crc kubenswrapper[4675]: E0219 09:02:23.635586 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635612 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon" Feb 19 09:02:23 crc kubenswrapper[4675]: E0219 09:02:23.635631 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon-log" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635652 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon-log" Feb 19 09:02:23 crc kubenswrapper[4675]: E0219 09:02:23.635667 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635673 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon" Feb 19 09:02:23 crc kubenswrapper[4675]: E0219 09:02:23.635701 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon-log" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635708 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon-log" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635891 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon-log" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635909 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635923 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" containerName="horizon" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.635932 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" containerName="horizon-log" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.636971 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.647855 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57f4c8679f-grtkm"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.663945 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7db6784559-bx2jh" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": read tcp 10.217.0.2:39450->10.217.0.157:9696: read: connection reset by peer" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.668029 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-ovndb-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.668087 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-combined-ca-bundle\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.668119 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-httpd-config\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.668201 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtldr\" (UniqueName: \"kubernetes.io/projected/81ba53aa-e359-412c-abe4-b91662a9996a-kube-api-access-dtldr\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.668270 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-internal-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.668380 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-config\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.668420 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-public-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.709474 4675 scope.go:117] "RemoveContainer" containerID="40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5" Feb 19 09:02:23 crc kubenswrapper[4675]: W0219 09:02:23.748588 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd5bc74e_534c_4ac6_a996_fdba94bad6d7.slice/crio-5def336c8b1901171db7b2f357c3caad11a81276f70b8e7f5ec5dbddff27e998 WatchSource:0}: Error finding container 5def336c8b1901171db7b2f357c3caad11a81276f70b8e7f5ec5dbddff27e998: Status 404 returned error can't find the container with id 5def336c8b1901171db7b2f357c3caad11a81276f70b8e7f5ec5dbddff27e998 Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.769811 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-ovndb-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.769900 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-combined-ca-bundle\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.769920 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-httpd-config\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.769971 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtldr\" (UniqueName: \"kubernetes.io/projected/81ba53aa-e359-412c-abe4-b91662a9996a-kube-api-access-dtldr\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.770010 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-internal-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.770068 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-config\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.770089 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-public-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.783955 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-public-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.785556 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-httpd-config\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.786939 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-ovndb-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.788839 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-combined-ca-bundle\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.790152 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-internal-tls-certs\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.790468 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtldr\" (UniqueName: \"kubernetes.io/projected/81ba53aa-e359-412c-abe4-b91662a9996a-kube-api-access-dtldr\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.799868 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/81ba53aa-e359-412c-abe4-b91662a9996a-config\") pod \"neutron-57f4c8679f-grtkm\" (UID: \"81ba53aa-e359-412c-abe4-b91662a9996a\") " pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.913892 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.953297 4675 scope.go:117] "RemoveContainer" containerID="006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad" Feb 19 09:02:23 crc kubenswrapper[4675]: E0219 09:02:23.954063 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad\": container with ID starting with 006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad not found: ID does not exist" containerID="006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.954102 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad"} err="failed to get container status \"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad\": rpc error: code = NotFound desc = could not find container \"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad\": container with ID starting with 006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad not found: ID does not exist" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.954134 4675 scope.go:117] "RemoveContainer" containerID="40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5" Feb 19 09:02:23 crc kubenswrapper[4675]: E0219 09:02:23.954808 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5\": container with ID starting with 40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5 not found: ID does not exist" containerID="40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.954859 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5"} err="failed to get container status \"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5\": rpc error: code = NotFound desc = could not find container \"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5\": container with ID starting with 40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5 not found: ID does not exist" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.954890 4675 scope.go:117] "RemoveContainer" containerID="006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.955536 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad"} err="failed to get container status \"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad\": rpc error: code = NotFound desc = could not find container \"006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad\": container with ID starting with 006a705661f5bc42c5ff7494cbbd5fa9ac7894df15765f73a17150534a1624ad not found: ID does not exist" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.955582 4675 scope.go:117] "RemoveContainer" containerID="40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.956018 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5"} err="failed to get container status \"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5\": rpc error: code = NotFound desc = could not find container \"40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5\": container with ID starting with 40a3710f1f5ec2ad66422e5bbefb447d407b4b7754eb773698f68f38dc70bee5 not found: ID does not exist" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.956041 4675 scope.go:117] "RemoveContainer" containerID="80ec74317bd11d88956bcf2d002aa1462a757c1812cf26956b46d048b2b76569" Feb 19 09:02:23 crc kubenswrapper[4675]: I0219 09:02:23.971552 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.231405 4675 scope.go:117] "RemoveContainer" containerID="6c1fd05e0d490ce48f6728f5ab42948031b2d6da4ff8ab7aba93446c7a42a968" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.421287 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.471954 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"457624d6-3a3d-406f-bc5e-626f1a0244ac","Type":"ContainerStarted","Data":"e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.501903 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fd5bc74e-534c-4ac6-a996-fdba94bad6d7","Type":"ContainerStarted","Data":"5def336c8b1901171db7b2f357c3caad11a81276f70b8e7f5ec5dbddff27e998"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.526885 4675 generic.go:334] "Generic (PLEG): container finished" podID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerID="5bd73ab1d57b0d778ca4fedc285167e3357cb43490ec9fc6d5ee18aa167b0dd2" exitCode=0 Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.526977 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" event={"ID":"825f8331-17ce-4e8e-beb5-4f2807fc4ceb","Type":"ContainerDied","Data":"5bd73ab1d57b0d778ca4fedc285167e3357cb43490ec9fc6d5ee18aa167b0dd2"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.563438 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-swift-storage-0\") pod \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.563561 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-sb\") pod \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.563613 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-config\") pod \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.563675 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n7nm\" (UniqueName: \"kubernetes.io/projected/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-kube-api-access-2n7nm\") pod \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.563735 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-svc\") pod \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.563752 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-nb\") pod \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\" (UID: \"1b94f207-4b1c-49ff-a2c2-ff15d94557a6\") " Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.568136 4675 generic.go:334] "Generic (PLEG): container finished" podID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerID="aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02" exitCode=0 Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.568248 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db6784559-bx2jh" event={"ID":"be374964-b557-48a9-b3db-68cf0e3d4a47","Type":"ContainerDied","Data":"aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.577386 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-kube-api-access-2n7nm" (OuterVolumeSpecName: "kube-api-access-2n7nm") pod "1b94f207-4b1c-49ff-a2c2-ff15d94557a6" (UID: "1b94f207-4b1c-49ff-a2c2-ff15d94557a6"). InnerVolumeSpecName "kube-api-access-2n7nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.591958 4675 generic.go:334] "Generic (PLEG): container finished" podID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerID="77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2" exitCode=0 Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.591996 4675 generic.go:334] "Generic (PLEG): container finished" podID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerID="48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5" exitCode=2 Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.592005 4675 generic.go:334] "Generic (PLEG): container finished" podID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerID="6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac" exitCode=0 Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.592075 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerDied","Data":"77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.592103 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerDied","Data":"48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.592113 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerDied","Data":"6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.602448 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b94f207-4b1c-49ff-a2c2-ff15d94557a6" (UID: "1b94f207-4b1c-49ff-a2c2-ff15d94557a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.602612 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-b2std" event={"ID":"1b94f207-4b1c-49ff-a2c2-ff15d94557a6","Type":"ContainerDied","Data":"e789b52065f282758d82e4a654ebec0192dd27b87a2075fec25ca429a6903571"} Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.602858 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-b2std" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.605308 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-config" (OuterVolumeSpecName: "config") pod "1b94f207-4b1c-49ff-a2c2-ff15d94557a6" (UID: "1b94f207-4b1c-49ff-a2c2-ff15d94557a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.609677 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b94f207-4b1c-49ff-a2c2-ff15d94557a6" (UID: "1b94f207-4b1c-49ff-a2c2-ff15d94557a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.624253 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1b94f207-4b1c-49ff-a2c2-ff15d94557a6" (UID: "1b94f207-4b1c-49ff-a2c2-ff15d94557a6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.639502 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b94f207-4b1c-49ff-a2c2-ff15d94557a6" (UID: "1b94f207-4b1c-49ff-a2c2-ff15d94557a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.666596 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.666771 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n7nm\" (UniqueName: \"kubernetes.io/projected/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-kube-api-access-2n7nm\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.666853 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.666949 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.667009 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.667063 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b94f207-4b1c-49ff-a2c2-ff15d94557a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.965370 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-b2std"] Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.974692 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-b2std"] Feb 19 09:02:24 crc kubenswrapper[4675]: I0219 09:02:24.987010 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57f4c8679f-grtkm"] Feb 19 09:02:25 crc kubenswrapper[4675]: I0219 09:02:25.121082 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f421816-e18a-4955-ba7a-557ed5e91ad3" path="/var/lib/kubelet/pods/0f421816-e18a-4955-ba7a-557ed5e91ad3/volumes" Feb 19 09:02:25 crc kubenswrapper[4675]: I0219 09:02:25.121775 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b94f207-4b1c-49ff-a2c2-ff15d94557a6" path="/var/lib/kubelet/pods/1b94f207-4b1c-49ff-a2c2-ff15d94557a6/volumes" Feb 19 09:02:25 crc kubenswrapper[4675]: I0219 09:02:25.122460 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e38966-737c-49a6-800a-3293f1fd65a9" path="/var/lib/kubelet/pods/d2e38966-737c-49a6-800a-3293f1fd65a9/volumes" Feb 19 09:02:25 crc kubenswrapper[4675]: I0219 09:02:25.300138 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:02:25 crc kubenswrapper[4675]: I0219 09:02:25.331268 4675 scope.go:117] "RemoveContainer" containerID="dd2513ddf92af792268c2797c706db1f07d9fc7ea2032eb00819672fe96c5b48" Feb 19 09:02:25 crc kubenswrapper[4675]: E0219 09:02:25.364788 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0824f353_528c_4847_acfd_8f8635d6102f.slice/crio-0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d\": RecentStats: unable to find data in memory cache]" Feb 19 09:02:25 crc kubenswrapper[4675]: I0219 09:02:25.615824 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57f4c8679f-grtkm" event={"ID":"81ba53aa-e359-412c-abe4-b91662a9996a","Type":"ContainerStarted","Data":"070e6247353237b5ea9e24496d98934190dd19b7621bdc44400375f70da20c01"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.463340 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.608916 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-log-httpd\") pod \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.609620 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-sg-core-conf-yaml\") pod \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.609740 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-scripts\") pod \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.609816 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl8hr\" (UniqueName: \"kubernetes.io/projected/ecee41ce-0a05-4ea5-8f91-81f760d1f093-kube-api-access-nl8hr\") pod \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.609910 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-config-data\") pod \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.610037 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-combined-ca-bundle\") pod \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.609805 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ecee41ce-0a05-4ea5-8f91-81f760d1f093" (UID: "ecee41ce-0a05-4ea5-8f91-81f760d1f093"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.610197 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-run-httpd\") pod \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\" (UID: \"ecee41ce-0a05-4ea5-8f91-81f760d1f093\") " Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.610706 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.611079 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ecee41ce-0a05-4ea5-8f91-81f760d1f093" (UID: "ecee41ce-0a05-4ea5-8f91-81f760d1f093"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.614182 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecee41ce-0a05-4ea5-8f91-81f760d1f093-kube-api-access-nl8hr" (OuterVolumeSpecName: "kube-api-access-nl8hr") pod "ecee41ce-0a05-4ea5-8f91-81f760d1f093" (UID: "ecee41ce-0a05-4ea5-8f91-81f760d1f093"). InnerVolumeSpecName "kube-api-access-nl8hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.616046 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-scripts" (OuterVolumeSpecName: "scripts") pod "ecee41ce-0a05-4ea5-8f91-81f760d1f093" (UID: "ecee41ce-0a05-4ea5-8f91-81f760d1f093"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.654954 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-bcb894f54-l299f" event={"ID":"8649bd53-643b-466d-a528-45d2aa799316","Type":"ContainerStarted","Data":"da00478f249a0af2f0fa76cdaf349954f84a88fb5b4768fb69b4a1542cbc5094"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.655034 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-bcb894f54-l299f" event={"ID":"8649bd53-643b-466d-a528-45d2aa799316","Type":"ContainerStarted","Data":"0917dc227c05b2b27667aad4d59c524aacaf57267a26e5219eb6b771e08202e0"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.656554 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ecee41ce-0a05-4ea5-8f91-81f760d1f093" (UID: "ecee41ce-0a05-4ea5-8f91-81f760d1f093"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.672422 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57f4c8679f-grtkm" event={"ID":"81ba53aa-e359-412c-abe4-b91662a9996a","Type":"ContainerStarted","Data":"c67a66a376f5daa9639c4191d4e10cfc3170eb4938aefa45668fe889aaa1b968"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.672499 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57f4c8679f-grtkm" event={"ID":"81ba53aa-e359-412c-abe4-b91662a9996a","Type":"ContainerStarted","Data":"f45ac816b6964762a2499d7046d2f42fab6c1c4e02909c3d6b50c4b2ed7bea8e"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.672933 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.683125 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7667dd7745-49ztz" event={"ID":"5a4a2232-762f-4fc1-83fd-45d33b408e5b","Type":"ContainerStarted","Data":"69b7db977d55a524954ce673ac37b23216d99768bdd4dc615016e36377091bd3"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.683192 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7667dd7745-49ztz" event={"ID":"5a4a2232-762f-4fc1-83fd-45d33b408e5b","Type":"ContainerStarted","Data":"f4531a0123211f1bfa08a37e6589be9a9e530af586f5f330da6b17f7689c4411"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.689753 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-bcb894f54-l299f" podStartSLOduration=9.369724761 podStartE2EDuration="12.689722381s" podCreationTimestamp="2026-02-19 09:02:14 +0000 UTC" firstStartedPulling="2026-02-19 09:02:22.223054271 +0000 UTC m=+1123.850144539" lastFinishedPulling="2026-02-19 09:02:25.543051891 +0000 UTC m=+1127.170142159" observedRunningTime="2026-02-19 09:02:26.685029395 +0000 UTC m=+1128.312119663" watchObservedRunningTime="2026-02-19 09:02:26.689722381 +0000 UTC m=+1128.316812659" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.694846 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"457624d6-3a3d-406f-bc5e-626f1a0244ac","Type":"ContainerStarted","Data":"3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.695091 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api-log" containerID="cri-o://e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6" gracePeriod=30 Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.695202 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.695242 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api" containerID="cri-o://3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136" gracePeriod=30 Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.705502 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fd5bc74e-534c-4ac6-a996-fdba94bad6d7","Type":"ContainerStarted","Data":"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.707880 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" event={"ID":"825f8331-17ce-4e8e-beb5-4f2807fc4ceb","Type":"ContainerStarted","Data":"1697cb8de82ab9deb11ca0300483af6c6ebaf528c52f0b463e5ba5235551636c"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.708937 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.713479 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecee41ce-0a05-4ea5-8f91-81f760d1f093-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.713508 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.713517 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.713528 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl8hr\" (UniqueName: \"kubernetes.io/projected/ecee41ce-0a05-4ea5-8f91-81f760d1f093-kube-api-access-nl8hr\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.727404 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecee41ce-0a05-4ea5-8f91-81f760d1f093" (UID: "ecee41ce-0a05-4ea5-8f91-81f760d1f093"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.730425 4675 generic.go:334] "Generic (PLEG): container finished" podID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerID="3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e" exitCode=0 Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.730471 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerDied","Data":"3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.730500 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecee41ce-0a05-4ea5-8f91-81f760d1f093","Type":"ContainerDied","Data":"234b9baa1d1bdbef293dd33d9a9502649b199a5f9dd800abec71534f272d17dc"} Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.730520 4675 scope.go:117] "RemoveContainer" containerID="77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.730728 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.742719 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7667dd7745-49ztz" podStartSLOduration=9.007949728 podStartE2EDuration="12.742690431s" podCreationTimestamp="2026-02-19 09:02:14 +0000 UTC" firstStartedPulling="2026-02-19 09:02:21.81284903 +0000 UTC m=+1123.439939298" lastFinishedPulling="2026-02-19 09:02:25.547589733 +0000 UTC m=+1127.174680001" observedRunningTime="2026-02-19 09:02:26.711424953 +0000 UTC m=+1128.338515221" watchObservedRunningTime="2026-02-19 09:02:26.742690431 +0000 UTC m=+1128.369780709" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.799070 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-config-data" (OuterVolumeSpecName: "config-data") pod "ecee41ce-0a05-4ea5-8f91-81f760d1f093" (UID: "ecee41ce-0a05-4ea5-8f91-81f760d1f093"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.799166 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-57f4c8679f-grtkm" podStartSLOduration=3.799141714 podStartE2EDuration="3.799141714s" podCreationTimestamp="2026-02-19 09:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:26.737226325 +0000 UTC m=+1128.364316593" watchObservedRunningTime="2026-02-19 09:02:26.799141714 +0000 UTC m=+1128.426231982" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.815621 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.816011 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecee41ce-0a05-4ea5-8f91-81f760d1f093-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.820530 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" podStartSLOduration=5.820499137 podStartE2EDuration="5.820499137s" podCreationTimestamp="2026-02-19 09:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:26.785273712 +0000 UTC m=+1128.412363980" watchObservedRunningTime="2026-02-19 09:02:26.820499137 +0000 UTC m=+1128.447589405" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.857329 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.857298944 podStartE2EDuration="5.857298944s" podCreationTimestamp="2026-02-19 09:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:26.818449932 +0000 UTC m=+1128.445540210" watchObservedRunningTime="2026-02-19 09:02:26.857298944 +0000 UTC m=+1128.484389212" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.882403 4675 scope.go:117] "RemoveContainer" containerID="48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.911178 4675 scope.go:117] "RemoveContainer" containerID="3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.958756 4675 scope.go:117] "RemoveContainer" containerID="6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.990307 4675 scope.go:117] "RemoveContainer" containerID="77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2" Feb 19 09:02:26 crc kubenswrapper[4675]: E0219 09:02:26.992457 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2\": container with ID starting with 77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2 not found: ID does not exist" containerID="77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.992504 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2"} err="failed to get container status \"77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2\": rpc error: code = NotFound desc = could not find container \"77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2\": container with ID starting with 77ab08837f18419494e05f0a49a090079a15a6861411566881802b018917c0f2 not found: ID does not exist" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.992559 4675 scope.go:117] "RemoveContainer" containerID="48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5" Feb 19 09:02:26 crc kubenswrapper[4675]: E0219 09:02:26.993026 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5\": container with ID starting with 48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5 not found: ID does not exist" containerID="48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.993076 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5"} err="failed to get container status \"48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5\": rpc error: code = NotFound desc = could not find container \"48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5\": container with ID starting with 48ed7ff3d1eedfdeb4a2fb1d3e2d0010dc6a1cd8943438951bb6a41a08e010b5 not found: ID does not exist" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.993099 4675 scope.go:117] "RemoveContainer" containerID="3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e" Feb 19 09:02:26 crc kubenswrapper[4675]: E0219 09:02:26.993462 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e\": container with ID starting with 3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e not found: ID does not exist" containerID="3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.993505 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e"} err="failed to get container status \"3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e\": rpc error: code = NotFound desc = could not find container \"3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e\": container with ID starting with 3b72dea84685a284e2cdfb895ce587c3561ce89c2d0c1b7b182ce6a643dae66e not found: ID does not exist" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.993519 4675 scope.go:117] "RemoveContainer" containerID="6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac" Feb 19 09:02:26 crc kubenswrapper[4675]: E0219 09:02:26.993964 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac\": container with ID starting with 6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac not found: ID does not exist" containerID="6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac" Feb 19 09:02:26 crc kubenswrapper[4675]: I0219 09:02:26.994009 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac"} err="failed to get container status \"6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac\": rpc error: code = NotFound desc = could not find container \"6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac\": container with ID starting with 6e2404c28e013329d31fa90e0a69bc49a5e1b1f56892929821ad2af2ea3eebac not found: ID does not exist" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.126780 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.134867 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.165543 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:27 crc kubenswrapper[4675]: E0219 09:02:27.165985 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-notification-agent" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166003 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-notification-agent" Feb 19 09:02:27 crc kubenswrapper[4675]: E0219 09:02:27.166015 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-central-agent" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166022 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-central-agent" Feb 19 09:02:27 crc kubenswrapper[4675]: E0219 09:02:27.166050 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="sg-core" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166057 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="sg-core" Feb 19 09:02:27 crc kubenswrapper[4675]: E0219 09:02:27.166074 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b94f207-4b1c-49ff-a2c2-ff15d94557a6" containerName="init" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166080 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b94f207-4b1c-49ff-a2c2-ff15d94557a6" containerName="init" Feb 19 09:02:27 crc kubenswrapper[4675]: E0219 09:02:27.166094 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="proxy-httpd" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166100 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="proxy-httpd" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166281 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-notification-agent" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166293 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="sg-core" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166304 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="proxy-httpd" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166323 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" containerName="ceilometer-central-agent" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.166332 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b94f207-4b1c-49ff-a2c2-ff15d94557a6" containerName="init" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.168110 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.171519 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.171711 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.186570 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.224657 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-log-httpd\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.224743 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.224830 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-run-httpd\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.224967 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-config-data\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.225107 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mtzb\" (UniqueName: \"kubernetes.io/projected/79aee500-9d06-41e6-8578-d92997ffbf71-kube-api-access-8mtzb\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.225172 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.225209 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-scripts\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.327381 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mtzb\" (UniqueName: \"kubernetes.io/projected/79aee500-9d06-41e6-8578-d92997ffbf71-kube-api-access-8mtzb\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.327451 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.327485 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-scripts\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.327564 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-log-httpd\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.327595 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.327667 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-run-httpd\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.327774 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-config-data\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.328210 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-log-httpd\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.329066 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-run-httpd\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.332956 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.334126 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.336739 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-scripts\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.347512 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-config-data\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.349788 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mtzb\" (UniqueName: \"kubernetes.io/projected/79aee500-9d06-41e6-8578-d92997ffbf71-kube-api-access-8mtzb\") pod \"ceilometer-0\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.541970 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.747766 4675 generic.go:334] "Generic (PLEG): container finished" podID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerID="e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6" exitCode=143 Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.747932 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"457624d6-3a3d-406f-bc5e-626f1a0244ac","Type":"ContainerDied","Data":"e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6"} Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.750191 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fd5bc74e-534c-4ac6-a996-fdba94bad6d7","Type":"ContainerStarted","Data":"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e"} Feb 19 09:02:27 crc kubenswrapper[4675]: I0219 09:02:27.786202 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.970096652 podStartE2EDuration="6.786183463s" podCreationTimestamp="2026-02-19 09:02:21 +0000 UTC" firstStartedPulling="2026-02-19 09:02:23.755352661 +0000 UTC m=+1125.382442929" lastFinishedPulling="2026-02-19 09:02:25.571439462 +0000 UTC m=+1127.198529740" observedRunningTime="2026-02-19 09:02:27.785397092 +0000 UTC m=+1129.412487360" watchObservedRunningTime="2026-02-19 09:02:27.786183463 +0000 UTC m=+1129.413273731" Feb 19 09:02:28 crc kubenswrapper[4675]: I0219 09:02:28.067434 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:28 crc kubenswrapper[4675]: W0219 09:02:28.072076 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79aee500_9d06_41e6_8578_d92997ffbf71.slice/crio-ef5c84f5e1a6dfbb9d7b487700d9493a9147ca063d085378387bbd4467907a24 WatchSource:0}: Error finding container ef5c84f5e1a6dfbb9d7b487700d9493a9147ca063d085378387bbd4467907a24: Status 404 returned error can't find the container with id ef5c84f5e1a6dfbb9d7b487700d9493a9147ca063d085378387bbd4467907a24 Feb 19 09:02:28 crc kubenswrapper[4675]: I0219 09:02:28.154880 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7db6784559-bx2jh" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": dial tcp 10.217.0.157:9696: connect: connection refused" Feb 19 09:02:28 crc kubenswrapper[4675]: I0219 09:02:28.778841 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerStarted","Data":"ef5c84f5e1a6dfbb9d7b487700d9493a9147ca063d085378387bbd4467907a24"} Feb 19 09:02:29 crc kubenswrapper[4675]: I0219 09:02:29.114727 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecee41ce-0a05-4ea5-8f91-81f760d1f093" path="/var/lib/kubelet/pods/ecee41ce-0a05-4ea5-8f91-81f760d1f093/volumes" Feb 19 09:02:29 crc kubenswrapper[4675]: I0219 09:02:29.789201 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerStarted","Data":"73e7eba373ec39d3342adc66b93eb4d50d5cfe1ed164de244b954fb85d55f5a3"} Feb 19 09:02:29 crc kubenswrapper[4675]: I0219 09:02:29.843215 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.185433 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b4d586cf4-9nm5g" Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.283299 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-77b588c78-htngd"] Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.283738 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api-log" containerID="cri-o://766445bb13fa4bb76aa03bc764fe6e0ed98768419a7a601565e2b59dac045d04" gracePeriod=30 Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.284554 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api" containerID="cri-o://113cf947778153c5ef2210ccd374f45ccfefe3b42bbb558eb51f3361c9d4dc43" gracePeriod=30 Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.290871 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": EOF" Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.291222 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": EOF" Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.800942 4675 generic.go:334] "Generic (PLEG): container finished" podID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerID="766445bb13fa4bb76aa03bc764fe6e0ed98768419a7a601565e2b59dac045d04" exitCode=143 Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.801091 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77b588c78-htngd" event={"ID":"e4ff5bfa-5771-4076-8be7-ce62be77af12","Type":"ContainerDied","Data":"766445bb13fa4bb76aa03bc764fe6e0ed98768419a7a601565e2b59dac045d04"} Feb 19 09:02:30 crc kubenswrapper[4675]: I0219 09:02:30.805436 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerStarted","Data":"d617b957c95eff4f88924001f60e8624de5dd96820cffa6f59bd40acac6cd4eb"} Feb 19 09:02:31 crc kubenswrapper[4675]: I0219 09:02:31.918052 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerStarted","Data":"3e8efc9e098404ab897054cb9681ad3aa8e7b9abeb2702609b1a607093d373d7"} Feb 19 09:02:31 crc kubenswrapper[4675]: I0219 09:02:31.984766 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.052670 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-vbwkt"] Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.052917 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" podUID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerName="dnsmasq-dns" containerID="cri-o://b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508" gracePeriod=10 Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.680184 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.706458 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.780349 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb48b\" (UniqueName: \"kubernetes.io/projected/1d0b3e27-6fb5-45da-b411-8e05d3119dff-kube-api-access-vb48b\") pod \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.780426 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-sb\") pod \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.780595 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-nb\") pod \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.780613 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-svc\") pod \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.780652 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-config\") pod \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.780728 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-swift-storage-0\") pod \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\" (UID: \"1d0b3e27-6fb5-45da-b411-8e05d3119dff\") " Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.798440 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d0b3e27-6fb5-45da-b411-8e05d3119dff-kube-api-access-vb48b" (OuterVolumeSpecName: "kube-api-access-vb48b") pod "1d0b3e27-6fb5-45da-b411-8e05d3119dff" (UID: "1d0b3e27-6fb5-45da-b411-8e05d3119dff"). InnerVolumeSpecName "kube-api-access-vb48b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.858768 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1d0b3e27-6fb5-45da-b411-8e05d3119dff" (UID: "1d0b3e27-6fb5-45da-b411-8e05d3119dff"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.864487 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-config" (OuterVolumeSpecName: "config") pod "1d0b3e27-6fb5-45da-b411-8e05d3119dff" (UID: "1d0b3e27-6fb5-45da-b411-8e05d3119dff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.876127 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1d0b3e27-6fb5-45da-b411-8e05d3119dff" (UID: "1d0b3e27-6fb5-45da-b411-8e05d3119dff"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.879193 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1d0b3e27-6fb5-45da-b411-8e05d3119dff" (UID: "1d0b3e27-6fb5-45da-b411-8e05d3119dff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.883403 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.883436 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.883448 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.883459 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb48b\" (UniqueName: \"kubernetes.io/projected/1d0b3e27-6fb5-45da-b411-8e05d3119dff-kube-api-access-vb48b\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.883469 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.938316 4675 generic.go:334] "Generic (PLEG): container finished" podID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerID="b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508" exitCode=0 Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.938358 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" event={"ID":"1d0b3e27-6fb5-45da-b411-8e05d3119dff","Type":"ContainerDied","Data":"b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508"} Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.938409 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" event={"ID":"1d0b3e27-6fb5-45da-b411-8e05d3119dff","Type":"ContainerDied","Data":"aec85350438df707cc5926e8c3094b9d6f0c92c6851206e3a40b20485044962d"} Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.938426 4675 scope.go:117] "RemoveContainer" containerID="b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.938571 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-vbwkt" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.943298 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1d0b3e27-6fb5-45da-b411-8e05d3119dff" (UID: "1d0b3e27-6fb5-45da-b411-8e05d3119dff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.966089 4675 scope.go:117] "RemoveContainer" containerID="611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.986034 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d0b3e27-6fb5-45da-b411-8e05d3119dff-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.992887 4675 scope.go:117] "RemoveContainer" containerID="b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508" Feb 19 09:02:32 crc kubenswrapper[4675]: E0219 09:02:32.993450 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508\": container with ID starting with b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508 not found: ID does not exist" containerID="b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.993539 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508"} err="failed to get container status \"b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508\": rpc error: code = NotFound desc = could not find container \"b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508\": container with ID starting with b2eeba195e7695e361068b613a39d34bed97afd89ba5d4e078c906ca2e5a1508 not found: ID does not exist" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.993617 4675 scope.go:117] "RemoveContainer" containerID="611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01" Feb 19 09:02:32 crc kubenswrapper[4675]: E0219 09:02:32.994070 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01\": container with ID starting with 611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01 not found: ID does not exist" containerID="611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01" Feb 19 09:02:32 crc kubenswrapper[4675]: I0219 09:02:32.994148 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01"} err="failed to get container status \"611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01\": rpc error: code = NotFound desc = could not find container \"611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01\": container with ID starting with 611adfb3eafe33c959774b37eb1e76571345e794fcf313a4c4ba73fccbb3fe01 not found: ID does not exist" Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.116997 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.174690 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.294887 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-vbwkt"] Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.308825 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-vbwkt"] Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.800177 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.913790 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-config\") pod \"be374964-b557-48a9-b3db-68cf0e3d4a47\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.914074 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-public-tls-certs\") pod \"be374964-b557-48a9-b3db-68cf0e3d4a47\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.914165 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-internal-tls-certs\") pod \"be374964-b557-48a9-b3db-68cf0e3d4a47\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.914302 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b679c\" (UniqueName: \"kubernetes.io/projected/be374964-b557-48a9-b3db-68cf0e3d4a47-kube-api-access-b679c\") pod \"be374964-b557-48a9-b3db-68cf0e3d4a47\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.914428 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-httpd-config\") pod \"be374964-b557-48a9-b3db-68cf0e3d4a47\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.914577 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-ovndb-tls-certs\") pod \"be374964-b557-48a9-b3db-68cf0e3d4a47\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.914695 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-combined-ca-bundle\") pod \"be374964-b557-48a9-b3db-68cf0e3d4a47\" (UID: \"be374964-b557-48a9-b3db-68cf0e3d4a47\") " Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.924343 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "be374964-b557-48a9-b3db-68cf0e3d4a47" (UID: "be374964-b557-48a9-b3db-68cf0e3d4a47"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:33 crc kubenswrapper[4675]: I0219 09:02:33.934828 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be374964-b557-48a9-b3db-68cf0e3d4a47-kube-api-access-b679c" (OuterVolumeSpecName: "kube-api-access-b679c") pod "be374964-b557-48a9-b3db-68cf0e3d4a47" (UID: "be374964-b557-48a9-b3db-68cf0e3d4a47"). InnerVolumeSpecName "kube-api-access-b679c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.017932 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b679c\" (UniqueName: \"kubernetes.io/projected/be374964-b557-48a9-b3db-68cf0e3d4a47-kube-api-access-b679c\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.019501 4675 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.035846 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be374964-b557-48a9-b3db-68cf0e3d4a47" (UID: "be374964-b557-48a9-b3db-68cf0e3d4a47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.045932 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerStarted","Data":"9bfb5b73fcd3e6e1e9cfaa4636e9602b07150c665dffcb944f0dad5b917e1a7c"} Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.046237 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.075459 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-config" (OuterVolumeSpecName: "config") pod "be374964-b557-48a9-b3db-68cf0e3d4a47" (UID: "be374964-b557-48a9-b3db-68cf0e3d4a47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.082942 4675 generic.go:334] "Generic (PLEG): container finished" podID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerID="bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28" exitCode=0 Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.083165 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="cinder-scheduler" containerID="cri-o://483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5" gracePeriod=30 Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.083534 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db6784559-bx2jh" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.083950 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db6784559-bx2jh" event={"ID":"be374964-b557-48a9-b3db-68cf0e3d4a47","Type":"ContainerDied","Data":"bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28"} Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.083975 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db6784559-bx2jh" event={"ID":"be374964-b557-48a9-b3db-68cf0e3d4a47","Type":"ContainerDied","Data":"5f1b56c4ce8e4faf94f514e437b648a56d325f8004ad928c609e9579136a88b8"} Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.083990 4675 scope.go:117] "RemoveContainer" containerID="aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.084269 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="probe" containerID="cri-o://cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e" gracePeriod=30 Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.120964 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.121682 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.127786 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be374964-b557-48a9-b3db-68cf0e3d4a47" (UID: "be374964-b557-48a9-b3db-68cf0e3d4a47"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.157043 4675 scope.go:117] "RemoveContainer" containerID="bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.164799 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "be374964-b557-48a9-b3db-68cf0e3d4a47" (UID: "be374964-b557-48a9-b3db-68cf0e3d4a47"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.191843 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be374964-b557-48a9-b3db-68cf0e3d4a47" (UID: "be374964-b557-48a9-b3db-68cf0e3d4a47"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.195236 4675 scope.go:117] "RemoveContainer" containerID="aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02" Feb 19 09:02:34 crc kubenswrapper[4675]: E0219 09:02:34.199684 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02\": container with ID starting with aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02 not found: ID does not exist" containerID="aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.199728 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02"} err="failed to get container status \"aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02\": rpc error: code = NotFound desc = could not find container \"aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02\": container with ID starting with aba582bfd5390d3384631836fc3a47d169e9a5332b803505d337f54857aafd02 not found: ID does not exist" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.199752 4675 scope.go:117] "RemoveContainer" containerID="bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28" Feb 19 09:02:34 crc kubenswrapper[4675]: E0219 09:02:34.203740 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28\": container with ID starting with bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28 not found: ID does not exist" containerID="bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.203773 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28"} err="failed to get container status \"bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28\": rpc error: code = NotFound desc = could not find container \"bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28\": container with ID starting with bf1e4f094de4c0cde90ddccb7bc480d2ae3c46f20c982fb0b092fd4533faef28 not found: ID does not exist" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.224856 4675 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.225844 4675 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.227092 4675 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be374964-b557-48a9-b3db-68cf0e3d4a47-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.242849 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.271755 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.158265897 podStartE2EDuration="7.271733893s" podCreationTimestamp="2026-02-19 09:02:27 +0000 UTC" firstStartedPulling="2026-02-19 09:02:28.075315997 +0000 UTC m=+1129.702406265" lastFinishedPulling="2026-02-19 09:02:33.188783993 +0000 UTC m=+1134.815874261" observedRunningTime="2026-02-19 09:02:34.071211275 +0000 UTC m=+1135.698301543" watchObservedRunningTime="2026-02-19 09:02:34.271733893 +0000 UTC m=+1135.898824161" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.391532 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-8594567c89-8xr7n" Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.432286 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7db6784559-bx2jh"] Feb 19 09:02:34 crc kubenswrapper[4675]: I0219 09:02:34.444186 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7db6784559-bx2jh"] Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.077781 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.078147 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.115397 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" path="/var/lib/kubelet/pods/1d0b3e27-6fb5-45da-b411-8e05d3119dff/volumes" Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.116124 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" path="/var/lib/kubelet/pods/be374964-b557-48a9-b3db-68cf0e3d4a47/volumes" Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.121887 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.297414 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7c56db696b-nwkrv" Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.637416 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:02:35 crc kubenswrapper[4675]: I0219 09:02:35.637873 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:02:35 crc kubenswrapper[4675]: E0219 09:02:35.791870 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0824f353_528c_4847_acfd_8f8635d6102f.slice/crio-0a9f2851f26e6bc2c71f7b5cf3329a1e62059911211b82c549fbb7aec2a6180d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd5bc74e_534c_4ac6_a996_fdba94bad6d7.slice/crio-conmon-483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5.scope\": RecentStats: unable to find data in memory cache]" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.054995 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.097175 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.117227 4675 generic.go:334] "Generic (PLEG): container finished" podID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerID="cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e" exitCode=0 Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.117257 4675 generic.go:334] "Generic (PLEG): container finished" podID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerID="483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5" exitCode=0 Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.117331 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fd5bc74e-534c-4ac6-a996-fdba94bad6d7","Type":"ContainerDied","Data":"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e"} Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.117384 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fd5bc74e-534c-4ac6-a996-fdba94bad6d7","Type":"ContainerDied","Data":"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5"} Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.117396 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fd5bc74e-534c-4ac6-a996-fdba94bad6d7","Type":"ContainerDied","Data":"5def336c8b1901171db7b2f357c3caad11a81276f70b8e7f5ec5dbddff27e998"} Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.117415 4675 scope.go:117] "RemoveContainer" containerID="cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.117709 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.153823 4675 scope.go:117] "RemoveContainer" containerID="483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.205324 4675 scope.go:117] "RemoveContainer" containerID="cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e" Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.206899 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e\": container with ID starting with cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e not found: ID does not exist" containerID="cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.206943 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e"} err="failed to get container status \"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e\": rpc error: code = NotFound desc = could not find container \"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e\": container with ID starting with cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e not found: ID does not exist" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.206972 4675 scope.go:117] "RemoveContainer" containerID="483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.207451 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data-custom\") pod \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.207486 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fzzn\" (UniqueName: \"kubernetes.io/projected/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-kube-api-access-7fzzn\") pod \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.207527 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-combined-ca-bundle\") pod \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.207784 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data\") pod \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.207824 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-scripts\") pod \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.207868 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-etc-machine-id\") pod \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\" (UID: \"fd5bc74e-534c-4ac6-a996-fdba94bad6d7\") " Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.209486 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fd5bc74e-534c-4ac6-a996-fdba94bad6d7" (UID: "fd5bc74e-534c-4ac6-a996-fdba94bad6d7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.209840 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5\": container with ID starting with 483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5 not found: ID does not exist" containerID="483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.209869 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5"} err="failed to get container status \"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5\": rpc error: code = NotFound desc = could not find container \"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5\": container with ID starting with 483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5 not found: ID does not exist" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.209888 4675 scope.go:117] "RemoveContainer" containerID="cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.210285 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e"} err="failed to get container status \"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e\": rpc error: code = NotFound desc = could not find container \"cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e\": container with ID starting with cd1ee430f3b6544efdd2670816f2776020bbd77a593db805e043ce12dd535a7e not found: ID does not exist" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.210317 4675 scope.go:117] "RemoveContainer" containerID="483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.213291 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5"} err="failed to get container status \"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5\": rpc error: code = NotFound desc = could not find container \"483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5\": container with ID starting with 483c0399c350f4f805930d90ddad73dd54ae0c0947f7e0fc280b21151cd7c1b5 not found: ID does not exist" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.216134 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-scripts" (OuterVolumeSpecName: "scripts") pod "fd5bc74e-534c-4ac6-a996-fdba94bad6d7" (UID: "fd5bc74e-534c-4ac6-a996-fdba94bad6d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.218995 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-kube-api-access-7fzzn" (OuterVolumeSpecName: "kube-api-access-7fzzn") pod "fd5bc74e-534c-4ac6-a996-fdba94bad6d7" (UID: "fd5bc74e-534c-4ac6-a996-fdba94bad6d7"). InnerVolumeSpecName "kube-api-access-7fzzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.219277 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fd5bc74e-534c-4ac6-a996-fdba94bad6d7" (UID: "fd5bc74e-534c-4ac6-a996-fdba94bad6d7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.310283 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.310596 4675 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.310727 4675 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.310827 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fzzn\" (UniqueName: \"kubernetes.io/projected/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-kube-api-access-7fzzn\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.337224 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data" (OuterVolumeSpecName: "config-data") pod "fd5bc74e-534c-4ac6-a996-fdba94bad6d7" (UID: "fd5bc74e-534c-4ac6-a996-fdba94bad6d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.399762 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd5bc74e-534c-4ac6-a996-fdba94bad6d7" (UID: "fd5bc74e-534c-4ac6-a996-fdba94bad6d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.413223 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.413518 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5bc74e-534c-4ac6-a996-fdba94bad6d7-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.460023 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.476894 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.494959 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.495353 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-httpd" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495369 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-httpd" Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.495380 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="cinder-scheduler" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495388 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="cinder-scheduler" Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.495397 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerName="dnsmasq-dns" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495403 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerName="dnsmasq-dns" Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.495424 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-api" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495430 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-api" Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.495439 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="probe" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495446 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="probe" Feb 19 09:02:36 crc kubenswrapper[4675]: E0219 09:02:36.495471 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerName="init" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495477 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerName="init" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495630 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d0b3e27-6fb5-45da-b411-8e05d3119dff" containerName="dnsmasq-dns" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495656 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="probe" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495667 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-api" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495678 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="be374964-b557-48a9-b3db-68cf0e3d4a47" containerName="neutron-httpd" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.495690 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" containerName="cinder-scheduler" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.496607 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.501261 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.521158 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.616673 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e85d2864-3bca-456e-a6a2-b5f4bc385b15-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.616987 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-config-data\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.617082 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.617156 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdz98\" (UniqueName: \"kubernetes.io/projected/e85d2864-3bca-456e-a6a2-b5f4bc385b15-kube-api-access-zdz98\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.617239 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-scripts\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.617557 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.719545 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-scripts\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.719697 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.719759 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e85d2864-3bca-456e-a6a2-b5f4bc385b15-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.719787 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-config-data\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.719821 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.719840 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdz98\" (UniqueName: \"kubernetes.io/projected/e85d2864-3bca-456e-a6a2-b5f4bc385b15-kube-api-access-zdz98\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.720055 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e85d2864-3bca-456e-a6a2-b5f4bc385b15-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.724473 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-scripts\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.724976 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.728238 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-config-data\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.730734 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e85d2864-3bca-456e-a6a2-b5f4bc385b15-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.746695 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdz98\" (UniqueName: \"kubernetes.io/projected/e85d2864-3bca-456e-a6a2-b5f4bc385b15-kube-api-access-zdz98\") pod \"cinder-scheduler-0\" (UID: \"e85d2864-3bca-456e-a6a2-b5f4bc385b15\") " pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.816404 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.846774 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-64d55559f4-zwfgc" Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.928792 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bbbd6c858-v6gv9"] Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.929162 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon-log" containerID="cri-o://56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff" gracePeriod=30 Feb 19 09:02:36 crc kubenswrapper[4675]: I0219 09:02:36.929399 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" containerID="cri-o://0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2" gracePeriod=30 Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.083168 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.084808 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.089037 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.089087 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.089349 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xgrcf" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.150578 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd5bc74e-534c-4ac6-a996-fdba94bad6d7" path="/var/lib/kubelet/pods/fd5bc74e-534c-4ac6-a996-fdba94bad6d7/volumes" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.151489 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.240832 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f27e5-af96-47cd-9550-dfa37033f584-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.240915 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhmf4\" (UniqueName: \"kubernetes.io/projected/2b0f27e5-af96-47cd-9550-dfa37033f584-kube-api-access-dhmf4\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.241116 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b0f27e5-af96-47cd-9550-dfa37033f584-openstack-config-secret\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.241176 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2b0f27e5-af96-47cd-9550-dfa37033f584-openstack-config\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.342934 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhmf4\" (UniqueName: \"kubernetes.io/projected/2b0f27e5-af96-47cd-9550-dfa37033f584-kube-api-access-dhmf4\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.343074 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b0f27e5-af96-47cd-9550-dfa37033f584-openstack-config-secret\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.343122 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2b0f27e5-af96-47cd-9550-dfa37033f584-openstack-config\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.343220 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f27e5-af96-47cd-9550-dfa37033f584-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.344484 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2b0f27e5-af96-47cd-9550-dfa37033f584-openstack-config\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.348209 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f27e5-af96-47cd-9550-dfa37033f584-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.349313 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2b0f27e5-af96-47cd-9550-dfa37033f584-openstack-config-secret\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.359857 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhmf4\" (UniqueName: \"kubernetes.io/projected/2b0f27e5-af96-47cd-9550-dfa37033f584-kube-api-access-dhmf4\") pod \"openstackclient\" (UID: \"2b0f27e5-af96-47cd-9550-dfa37033f584\") " pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.423930 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.426804 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 19 09:02:37 crc kubenswrapper[4675]: W0219 09:02:37.427163 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode85d2864_3bca_456e_a6a2_b5f4bc385b15.slice/crio-781a2a19721d8efc7b350415c4fda26aea7fe8c77b3b3af73a7736f95444a595 WatchSource:0}: Error finding container 781a2a19721d8efc7b350415c4fda26aea7fe8c77b3b3af73a7736f95444a595: Status 404 returned error can't find the container with id 781a2a19721d8efc7b350415c4fda26aea7fe8c77b3b3af73a7736f95444a595 Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.808397 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:55914->10.217.0.165:9311: read: connection reset by peer" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.809209 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77b588c78-htngd" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:55912->10.217.0.165:9311: read: connection reset by peer" Feb 19 09:02:37 crc kubenswrapper[4675]: I0219 09:02:37.976757 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.185066 4675 generic.go:334] "Generic (PLEG): container finished" podID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerID="113cf947778153c5ef2210ccd374f45ccfefe3b42bbb558eb51f3361c9d4dc43" exitCode=0 Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.185143 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77b588c78-htngd" event={"ID":"e4ff5bfa-5771-4076-8be7-ce62be77af12","Type":"ContainerDied","Data":"113cf947778153c5ef2210ccd374f45ccfefe3b42bbb558eb51f3361c9d4dc43"} Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.202397 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e85d2864-3bca-456e-a6a2-b5f4bc385b15","Type":"ContainerStarted","Data":"781a2a19721d8efc7b350415c4fda26aea7fe8c77b3b3af73a7736f95444a595"} Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.204015 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2b0f27e5-af96-47cd-9550-dfa37033f584","Type":"ContainerStarted","Data":"ba7c6ebb79bdd768769eb6b2358670680b52d4e4c90f9840e5dc1cc78f81be24"} Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.382008 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.467379 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ff5bfa-5771-4076-8be7-ce62be77af12-logs\") pod \"e4ff5bfa-5771-4076-8be7-ce62be77af12\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.467591 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data-custom\") pod \"e4ff5bfa-5771-4076-8be7-ce62be77af12\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.467693 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-combined-ca-bundle\") pod \"e4ff5bfa-5771-4076-8be7-ce62be77af12\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.467775 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs9wb\" (UniqueName: \"kubernetes.io/projected/e4ff5bfa-5771-4076-8be7-ce62be77af12-kube-api-access-hs9wb\") pod \"e4ff5bfa-5771-4076-8be7-ce62be77af12\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.467807 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data\") pod \"e4ff5bfa-5771-4076-8be7-ce62be77af12\" (UID: \"e4ff5bfa-5771-4076-8be7-ce62be77af12\") " Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.468863 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4ff5bfa-5771-4076-8be7-ce62be77af12-logs" (OuterVolumeSpecName: "logs") pod "e4ff5bfa-5771-4076-8be7-ce62be77af12" (UID: "e4ff5bfa-5771-4076-8be7-ce62be77af12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.473674 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ff5bfa-5771-4076-8be7-ce62be77af12-kube-api-access-hs9wb" (OuterVolumeSpecName: "kube-api-access-hs9wb") pod "e4ff5bfa-5771-4076-8be7-ce62be77af12" (UID: "e4ff5bfa-5771-4076-8be7-ce62be77af12"). InnerVolumeSpecName "kube-api-access-hs9wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.474820 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e4ff5bfa-5771-4076-8be7-ce62be77af12" (UID: "e4ff5bfa-5771-4076-8be7-ce62be77af12"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.505895 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4ff5bfa-5771-4076-8be7-ce62be77af12" (UID: "e4ff5bfa-5771-4076-8be7-ce62be77af12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.557256 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data" (OuterVolumeSpecName: "config-data") pod "e4ff5bfa-5771-4076-8be7-ce62be77af12" (UID: "e4ff5bfa-5771-4076-8be7-ce62be77af12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.569498 4675 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.569544 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.569555 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs9wb\" (UniqueName: \"kubernetes.io/projected/e4ff5bfa-5771-4076-8be7-ce62be77af12-kube-api-access-hs9wb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.569568 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4ff5bfa-5771-4076-8be7-ce62be77af12-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:38 crc kubenswrapper[4675]: I0219 09:02:38.569577 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4ff5bfa-5771-4076-8be7-ce62be77af12-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.218531 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77b588c78-htngd" event={"ID":"e4ff5bfa-5771-4076-8be7-ce62be77af12","Type":"ContainerDied","Data":"fde35d31b251aadfcc817ebe17c59402316931657d2603038661d4b66a879a03"} Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.219116 4675 scope.go:117] "RemoveContainer" containerID="113cf947778153c5ef2210ccd374f45ccfefe3b42bbb558eb51f3361c9d4dc43" Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.219323 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77b588c78-htngd" Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.227024 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e85d2864-3bca-456e-a6a2-b5f4bc385b15","Type":"ContainerStarted","Data":"00a220a714a15e0d43b12762b637ca55174cfc11f4c454c7f5a7abd92ce4290e"} Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.227079 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e85d2864-3bca-456e-a6a2-b5f4bc385b15","Type":"ContainerStarted","Data":"2577a398e94617951727edd29304317d41d4819c64941c64d7e68c7539455285"} Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.261704 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-77b588c78-htngd"] Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.273458 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-77b588c78-htngd"] Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.290183 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.290133929 podStartE2EDuration="3.290133929s" podCreationTimestamp="2026-02-19 09:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:39.268727365 +0000 UTC m=+1140.895817623" watchObservedRunningTime="2026-02-19 09:02:39.290133929 +0000 UTC m=+1140.917224197" Feb 19 09:02:39 crc kubenswrapper[4675]: I0219 09:02:39.302054 4675 scope.go:117] "RemoveContainer" containerID="766445bb13fa4bb76aa03bc764fe6e0ed98768419a7a601565e2b59dac045d04" Feb 19 09:02:40 crc kubenswrapper[4675]: I0219 09:02:40.115510 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:42832->10.217.0.149:8443: read: connection reset by peer" Feb 19 09:02:40 crc kubenswrapper[4675]: I0219 09:02:40.244533 4675 generic.go:334] "Generic (PLEG): container finished" podID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerID="0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2" exitCode=0 Feb 19 09:02:40 crc kubenswrapper[4675]: I0219 09:02:40.244658 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bbbd6c858-v6gv9" event={"ID":"d15338c6-aba7-4bec-a9c6-293e10903a7b","Type":"ContainerDied","Data":"0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2"} Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.029030 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-676ddcfdc-554zf"] Feb 19 09:02:41 crc kubenswrapper[4675]: E0219 09:02:41.029436 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.029459 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api" Feb 19 09:02:41 crc kubenswrapper[4675]: E0219 09:02:41.029473 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api-log" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.029479 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api-log" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.029683 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api-log" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.029710 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" containerName="barbican-api" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.030701 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.033363 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.033564 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.033608 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.043100 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-676ddcfdc-554zf"] Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.119968 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ff5bfa-5771-4076-8be7-ce62be77af12" path="/var/lib/kubelet/pods/e4ff5bfa-5771-4076-8be7-ce62be77af12/volumes" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125566 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c331033a-dd3e-4871-8e7d-85081578aace-log-httpd\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125650 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c331033a-dd3e-4871-8e7d-85081578aace-etc-swift\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125718 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-combined-ca-bundle\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125737 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-public-tls-certs\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125758 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-config-data\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125778 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c331033a-dd3e-4871-8e7d-85081578aace-run-httpd\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125802 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4wft\" (UniqueName: \"kubernetes.io/projected/c331033a-dd3e-4871-8e7d-85081578aace-kube-api-access-t4wft\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.125825 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-internal-tls-certs\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.227790 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-combined-ca-bundle\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.227831 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-public-tls-certs\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.227869 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-config-data\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.227892 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c331033a-dd3e-4871-8e7d-85081578aace-run-httpd\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.229232 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4wft\" (UniqueName: \"kubernetes.io/projected/c331033a-dd3e-4871-8e7d-85081578aace-kube-api-access-t4wft\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.229300 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-internal-tls-certs\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.229044 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c331033a-dd3e-4871-8e7d-85081578aace-run-httpd\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.229487 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c331033a-dd3e-4871-8e7d-85081578aace-log-httpd\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.229542 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c331033a-dd3e-4871-8e7d-85081578aace-etc-swift\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.229971 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c331033a-dd3e-4871-8e7d-85081578aace-log-httpd\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.234870 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-public-tls-certs\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.238361 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-combined-ca-bundle\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.238993 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-internal-tls-certs\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.240105 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c331033a-dd3e-4871-8e7d-85081578aace-etc-swift\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.249002 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c331033a-dd3e-4871-8e7d-85081578aace-config-data\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.251474 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4wft\" (UniqueName: \"kubernetes.io/projected/c331033a-dd3e-4871-8e7d-85081578aace-kube-api-access-t4wft\") pod \"swift-proxy-676ddcfdc-554zf\" (UID: \"c331033a-dd3e-4871-8e7d-85081578aace\") " pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.371402 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:41 crc kubenswrapper[4675]: I0219 09:02:41.816732 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 19 09:02:42 crc kubenswrapper[4675]: I0219 09:02:42.003785 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-676ddcfdc-554zf"] Feb 19 09:02:42 crc kubenswrapper[4675]: I0219 09:02:42.285135 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-676ddcfdc-554zf" event={"ID":"c331033a-dd3e-4871-8e7d-85081578aace","Type":"ContainerStarted","Data":"250a2bec97b027697f54cd7de3d4c79da463f8d948625a053975ea0a7b34e73b"} Feb 19 09:02:43 crc kubenswrapper[4675]: I0219 09:02:43.319541 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-676ddcfdc-554zf" event={"ID":"c331033a-dd3e-4871-8e7d-85081578aace","Type":"ContainerStarted","Data":"7aecc401c677350330e03e52e7aa5c535439f715b98632567bc3a3d3eefbe2a1"} Feb 19 09:02:43 crc kubenswrapper[4675]: I0219 09:02:43.321000 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:43 crc kubenswrapper[4675]: I0219 09:02:43.321091 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-676ddcfdc-554zf" event={"ID":"c331033a-dd3e-4871-8e7d-85081578aace","Type":"ContainerStarted","Data":"0625e5b7a6b946b5286d446f85095351ff1fc06f6f0924fedcfc66fbee1a57c7"} Feb 19 09:02:43 crc kubenswrapper[4675]: I0219 09:02:43.321177 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:44 crc kubenswrapper[4675]: I0219 09:02:44.270731 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-676ddcfdc-554zf" podStartSLOduration=3.270708501 podStartE2EDuration="3.270708501s" podCreationTimestamp="2026-02-19 09:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:43.354989804 +0000 UTC m=+1144.982080072" watchObservedRunningTime="2026-02-19 09:02:44.270708501 +0000 UTC m=+1145.897798769" Feb 19 09:02:44 crc kubenswrapper[4675]: I0219 09:02:44.282733 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:44 crc kubenswrapper[4675]: I0219 09:02:44.283238 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-central-agent" containerID="cri-o://73e7eba373ec39d3342adc66b93eb4d50d5cfe1ed164de244b954fb85d55f5a3" gracePeriod=30 Feb 19 09:02:44 crc kubenswrapper[4675]: I0219 09:02:44.283333 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="sg-core" containerID="cri-o://3e8efc9e098404ab897054cb9681ad3aa8e7b9abeb2702609b1a607093d373d7" gracePeriod=30 Feb 19 09:02:44 crc kubenswrapper[4675]: I0219 09:02:44.283466 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-notification-agent" containerID="cri-o://d617b957c95eff4f88924001f60e8624de5dd96820cffa6f59bd40acac6cd4eb" gracePeriod=30 Feb 19 09:02:44 crc kubenswrapper[4675]: I0219 09:02:44.283497 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="proxy-httpd" containerID="cri-o://9bfb5b73fcd3e6e1e9cfaa4636e9602b07150c665dffcb944f0dad5b917e1a7c" gracePeriod=30 Feb 19 09:02:44 crc kubenswrapper[4675]: I0219 09:02:44.302156 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342126 4675 generic.go:334] "Generic (PLEG): container finished" podID="79aee500-9d06-41e6-8578-d92997ffbf71" containerID="9bfb5b73fcd3e6e1e9cfaa4636e9602b07150c665dffcb944f0dad5b917e1a7c" exitCode=0 Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342164 4675 generic.go:334] "Generic (PLEG): container finished" podID="79aee500-9d06-41e6-8578-d92997ffbf71" containerID="3e8efc9e098404ab897054cb9681ad3aa8e7b9abeb2702609b1a607093d373d7" exitCode=2 Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342177 4675 generic.go:334] "Generic (PLEG): container finished" podID="79aee500-9d06-41e6-8578-d92997ffbf71" containerID="d617b957c95eff4f88924001f60e8624de5dd96820cffa6f59bd40acac6cd4eb" exitCode=0 Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342187 4675 generic.go:334] "Generic (PLEG): container finished" podID="79aee500-9d06-41e6-8578-d92997ffbf71" containerID="73e7eba373ec39d3342adc66b93eb4d50d5cfe1ed164de244b954fb85d55f5a3" exitCode=0 Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342200 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerDied","Data":"9bfb5b73fcd3e6e1e9cfaa4636e9602b07150c665dffcb944f0dad5b917e1a7c"} Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342246 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerDied","Data":"3e8efc9e098404ab897054cb9681ad3aa8e7b9abeb2702609b1a607093d373d7"} Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342256 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerDied","Data":"d617b957c95eff4f88924001f60e8624de5dd96820cffa6f59bd40acac6cd4eb"} Feb 19 09:02:45 crc kubenswrapper[4675]: I0219 09:02:45.342265 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerDied","Data":"73e7eba373ec39d3342adc66b93eb4d50d5cfe1ed164de244b954fb85d55f5a3"} Feb 19 09:02:47 crc kubenswrapper[4675]: I0219 09:02:47.061285 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 19 09:02:48 crc kubenswrapper[4675]: I0219 09:02:48.918072 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:48 crc kubenswrapper[4675]: I0219 09:02:48.918748 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-log" containerID="cri-o://c966f779a7de37b0c7a32b6548053926814c76c76852307d77a6619f202d603a" gracePeriod=30 Feb 19 09:02:48 crc kubenswrapper[4675]: I0219 09:02:48.919170 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-httpd" containerID="cri-o://4491d6009caa431167f30a62aef780ddc463bf8377c2eda805101714b027a37a" gracePeriod=30 Feb 19 09:02:49 crc kubenswrapper[4675]: I0219 09:02:49.392771 4675 generic.go:334] "Generic (PLEG): container finished" podID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerID="c966f779a7de37b0c7a32b6548053926814c76c76852307d77a6619f202d603a" exitCode=143 Feb 19 09:02:49 crc kubenswrapper[4675]: I0219 09:02:49.392816 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c15d3213-8b81-4da3-851b-a6927b51f7aa","Type":"ContainerDied","Data":"c966f779a7de37b0c7a32b6548053926814c76c76852307d77a6619f202d603a"} Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.057356 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.058721 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.642500 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.754549 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mtzb\" (UniqueName: \"kubernetes.io/projected/79aee500-9d06-41e6-8578-d92997ffbf71-kube-api-access-8mtzb\") pod \"79aee500-9d06-41e6-8578-d92997ffbf71\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.754883 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-config-data\") pod \"79aee500-9d06-41e6-8578-d92997ffbf71\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.755066 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-log-httpd\") pod \"79aee500-9d06-41e6-8578-d92997ffbf71\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.755102 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-scripts\") pod \"79aee500-9d06-41e6-8578-d92997ffbf71\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.755149 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-sg-core-conf-yaml\") pod \"79aee500-9d06-41e6-8578-d92997ffbf71\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.755185 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-run-httpd\") pod \"79aee500-9d06-41e6-8578-d92997ffbf71\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.755250 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-combined-ca-bundle\") pod \"79aee500-9d06-41e6-8578-d92997ffbf71\" (UID: \"79aee500-9d06-41e6-8578-d92997ffbf71\") " Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.755912 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "79aee500-9d06-41e6-8578-d92997ffbf71" (UID: "79aee500-9d06-41e6-8578-d92997ffbf71"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.756070 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "79aee500-9d06-41e6-8578-d92997ffbf71" (UID: "79aee500-9d06-41e6-8578-d92997ffbf71"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.760526 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-scripts" (OuterVolumeSpecName: "scripts") pod "79aee500-9d06-41e6-8578-d92997ffbf71" (UID: "79aee500-9d06-41e6-8578-d92997ffbf71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.760804 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79aee500-9d06-41e6-8578-d92997ffbf71-kube-api-access-8mtzb" (OuterVolumeSpecName: "kube-api-access-8mtzb") pod "79aee500-9d06-41e6-8578-d92997ffbf71" (UID: "79aee500-9d06-41e6-8578-d92997ffbf71"). InnerVolumeSpecName "kube-api-access-8mtzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.790098 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "79aee500-9d06-41e6-8578-d92997ffbf71" (UID: "79aee500-9d06-41e6-8578-d92997ffbf71"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.832114 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79aee500-9d06-41e6-8578-d92997ffbf71" (UID: "79aee500-9d06-41e6-8578-d92997ffbf71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.848845 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-config-data" (OuterVolumeSpecName: "config-data") pod "79aee500-9d06-41e6-8578-d92997ffbf71" (UID: "79aee500-9d06-41e6-8578-d92997ffbf71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.857720 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.857750 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.857761 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.857775 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79aee500-9d06-41e6-8578-d92997ffbf71-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.857786 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.857799 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mtzb\" (UniqueName: \"kubernetes.io/projected/79aee500-9d06-41e6-8578-d92997ffbf71-kube-api-access-8mtzb\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:50 crc kubenswrapper[4675]: I0219 09:02:50.857812 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79aee500-9d06-41e6-8578-d92997ffbf71-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.381879 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.388882 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-676ddcfdc-554zf" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.435234 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79aee500-9d06-41e6-8578-d92997ffbf71","Type":"ContainerDied","Data":"ef5c84f5e1a6dfbb9d7b487700d9493a9147ca063d085378387bbd4467907a24"} Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.435262 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.435301 4675 scope.go:117] "RemoveContainer" containerID="9bfb5b73fcd3e6e1e9cfaa4636e9602b07150c665dffcb944f0dad5b917e1a7c" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.439149 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2b0f27e5-af96-47cd-9550-dfa37033f584","Type":"ContainerStarted","Data":"92591abe445f724cbcd0c242ca1d860530673db3e614e6399193fecace760e95"} Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.477787 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.478864 4675 scope.go:117] "RemoveContainer" containerID="3e8efc9e098404ab897054cb9681ad3aa8e7b9abeb2702609b1a607093d373d7" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.498217 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.508613 4675 scope.go:117] "RemoveContainer" containerID="d617b957c95eff4f88924001f60e8624de5dd96820cffa6f59bd40acac6cd4eb" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.514670 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:51 crc kubenswrapper[4675]: E0219 09:02:51.515091 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-central-agent" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515104 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-central-agent" Feb 19 09:02:51 crc kubenswrapper[4675]: E0219 09:02:51.515123 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-notification-agent" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515130 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-notification-agent" Feb 19 09:02:51 crc kubenswrapper[4675]: E0219 09:02:51.515146 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="proxy-httpd" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515153 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="proxy-httpd" Feb 19 09:02:51 crc kubenswrapper[4675]: E0219 09:02:51.515171 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="sg-core" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515176 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="sg-core" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515343 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-notification-agent" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515365 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="proxy-httpd" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515377 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="ceilometer-central-agent" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.515386 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" containerName="sg-core" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.517014 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.522675 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.156518406 podStartE2EDuration="14.522654112s" podCreationTimestamp="2026-02-19 09:02:37 +0000 UTC" firstStartedPulling="2026-02-19 09:02:37.998203944 +0000 UTC m=+1139.625294212" lastFinishedPulling="2026-02-19 09:02:50.36433965 +0000 UTC m=+1151.991429918" observedRunningTime="2026-02-19 09:02:51.493234713 +0000 UTC m=+1153.120324981" watchObservedRunningTime="2026-02-19 09:02:51.522654112 +0000 UTC m=+1153.149744380" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.522744 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.522940 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.543872 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.546349 4675 scope.go:117] "RemoveContainer" containerID="73e7eba373ec39d3342adc66b93eb4d50d5cfe1ed164de244b954fb85d55f5a3" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.676009 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.676219 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-log-httpd\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.676262 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-config-data\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.676345 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk9dr\" (UniqueName: \"kubernetes.io/projected/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-kube-api-access-tk9dr\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.676481 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-scripts\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.676562 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.676737 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-run-httpd\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.779059 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-run-httpd\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.779150 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.779216 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-log-httpd\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.779243 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-config-data\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.779282 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk9dr\" (UniqueName: \"kubernetes.io/projected/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-kube-api-access-tk9dr\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.779341 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-scripts\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.779379 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.780819 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-run-httpd\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.780970 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-log-httpd\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.787782 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.788255 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-config-data\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.791445 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.797297 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-scripts\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.807486 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk9dr\" (UniqueName: \"kubernetes.io/projected/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-kube-api-access-tk9dr\") pod \"ceilometer-0\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " pod="openstack/ceilometer-0" Feb 19 09:02:51 crc kubenswrapper[4675]: I0219 09:02:51.833416 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.011060 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-crs9q"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.012312 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.021794 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-crs9q"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.193963 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-operator-scripts\") pod \"nova-api-db-create-crs9q\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.194072 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxc9g\" (UniqueName: \"kubernetes.io/projected/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-kube-api-access-gxc9g\") pod \"nova-api-db-create-crs9q\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.198030 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vk5pq"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.199869 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.214869 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-e4fa-account-create-update-rbg2p"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.219051 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.222975 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.223359 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vk5pq"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.243035 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e4fa-account-create-update-rbg2p"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.308183 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sbbg\" (UniqueName: \"kubernetes.io/projected/8b190a6d-42fd-487a-a90d-423a63e23603-kube-api-access-2sbbg\") pod \"nova-api-e4fa-account-create-update-rbg2p\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.308233 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc9wh\" (UniqueName: \"kubernetes.io/projected/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-kube-api-access-dc9wh\") pod \"nova-cell0-db-create-vk5pq\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.308274 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxc9g\" (UniqueName: \"kubernetes.io/projected/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-kube-api-access-gxc9g\") pod \"nova-api-db-create-crs9q\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.308328 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-operator-scripts\") pod \"nova-cell0-db-create-vk5pq\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.308400 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-operator-scripts\") pod \"nova-api-db-create-crs9q\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.308442 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b190a6d-42fd-487a-a90d-423a63e23603-operator-scripts\") pod \"nova-api-e4fa-account-create-update-rbg2p\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.309529 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-operator-scripts\") pod \"nova-api-db-create-crs9q\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.310610 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-q2tv7"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.320940 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.343515 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-q2tv7"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.347769 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxc9g\" (UniqueName: \"kubernetes.io/projected/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-kube-api-access-gxc9g\") pod \"nova-api-db-create-crs9q\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: W0219 09:02:52.352623 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5633bc1_203a_4dd9_9089_ee1e7a5a115e.slice/crio-dea5c963c7d98038602e67b43efb3d753775dbd0ddf0aec511a29054c5ccfbe0 WatchSource:0}: Error finding container dea5c963c7d98038602e67b43efb3d753775dbd0ddf0aec511a29054c5ccfbe0: Status 404 returned error can't find the container with id dea5c963c7d98038602e67b43efb3d753775dbd0ddf0aec511a29054c5ccfbe0 Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.380860 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.404520 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0a16-account-create-update-n7pz4"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.405999 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.407997 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.409825 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxhtn\" (UniqueName: \"kubernetes.io/projected/461c89f2-a273-4964-b6fa-e6a858c7b97c-kube-api-access-vxhtn\") pod \"nova-cell1-db-create-q2tv7\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.409933 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b190a6d-42fd-487a-a90d-423a63e23603-operator-scripts\") pod \"nova-api-e4fa-account-create-update-rbg2p\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.409988 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/461c89f2-a273-4964-b6fa-e6a858c7b97c-operator-scripts\") pod \"nova-cell1-db-create-q2tv7\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.410022 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sbbg\" (UniqueName: \"kubernetes.io/projected/8b190a6d-42fd-487a-a90d-423a63e23603-kube-api-access-2sbbg\") pod \"nova-api-e4fa-account-create-update-rbg2p\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.410044 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc9wh\" (UniqueName: \"kubernetes.io/projected/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-kube-api-access-dc9wh\") pod \"nova-cell0-db-create-vk5pq\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.410089 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-operator-scripts\") pod \"nova-cell0-db-create-vk5pq\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.410925 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-operator-scripts\") pod \"nova-cell0-db-create-vk5pq\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.418277 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0a16-account-create-update-n7pz4"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.419914 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b190a6d-42fd-487a-a90d-423a63e23603-operator-scripts\") pod \"nova-api-e4fa-account-create-update-rbg2p\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.432904 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sbbg\" (UniqueName: \"kubernetes.io/projected/8b190a6d-42fd-487a-a90d-423a63e23603-kube-api-access-2sbbg\") pod \"nova-api-e4fa-account-create-update-rbg2p\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.434078 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc9wh\" (UniqueName: \"kubernetes.io/projected/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-kube-api-access-dc9wh\") pod \"nova-cell0-db-create-vk5pq\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.450392 4675 generic.go:334] "Generic (PLEG): container finished" podID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerID="4491d6009caa431167f30a62aef780ddc463bf8377c2eda805101714b027a37a" exitCode=0 Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.450468 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c15d3213-8b81-4da3-851b-a6927b51f7aa","Type":"ContainerDied","Data":"4491d6009caa431167f30a62aef780ddc463bf8377c2eda805101714b027a37a"} Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.452357 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerStarted","Data":"dea5c963c7d98038602e67b43efb3d753775dbd0ddf0aec511a29054c5ccfbe0"} Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.517099 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsp42\" (UniqueName: \"kubernetes.io/projected/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-kube-api-access-bsp42\") pod \"nova-cell0-0a16-account-create-update-n7pz4\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.517166 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxhtn\" (UniqueName: \"kubernetes.io/projected/461c89f2-a273-4964-b6fa-e6a858c7b97c-kube-api-access-vxhtn\") pod \"nova-cell1-db-create-q2tv7\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.517254 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-operator-scripts\") pod \"nova-cell0-0a16-account-create-update-n7pz4\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.517305 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/461c89f2-a273-4964-b6fa-e6a858c7b97c-operator-scripts\") pod \"nova-cell1-db-create-q2tv7\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.518388 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/461c89f2-a273-4964-b6fa-e6a858c7b97c-operator-scripts\") pod \"nova-cell1-db-create-q2tv7\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.537365 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxhtn\" (UniqueName: \"kubernetes.io/projected/461c89f2-a273-4964-b6fa-e6a858c7b97c-kube-api-access-vxhtn\") pod \"nova-cell1-db-create-q2tv7\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.583452 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.607660 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-61a4-account-create-update-rj4bg"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.609901 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.610566 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.616682 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.618575 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-61a4-account-create-update-rj4bg"] Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.618965 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-operator-scripts\") pod \"nova-cell0-0a16-account-create-update-n7pz4\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.619113 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsp42\" (UniqueName: \"kubernetes.io/projected/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-kube-api-access-bsp42\") pod \"nova-cell0-0a16-account-create-update-n7pz4\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.622503 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-operator-scripts\") pod \"nova-cell0-0a16-account-create-update-n7pz4\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.643914 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.662343 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsp42\" (UniqueName: \"kubernetes.io/projected/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-kube-api-access-bsp42\") pod \"nova-cell0-0a16-account-create-update-n7pz4\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.723818 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/179577e4-e654-46ef-b92e-47891c2ff281-operator-scripts\") pod \"nova-cell1-61a4-account-create-update-rj4bg\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.724142 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62qxs\" (UniqueName: \"kubernetes.io/projected/179577e4-e654-46ef-b92e-47891c2ff281-kube-api-access-62qxs\") pod \"nova-cell1-61a4-account-create-update-rj4bg\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.727327 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.794270 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.812231 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.825945 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/179577e4-e654-46ef-b92e-47891c2ff281-operator-scripts\") pod \"nova-cell1-61a4-account-create-update-rj4bg\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.825995 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62qxs\" (UniqueName: \"kubernetes.io/projected/179577e4-e654-46ef-b92e-47891c2ff281-kube-api-access-62qxs\") pod \"nova-cell1-61a4-account-create-update-rj4bg\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.827210 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/179577e4-e654-46ef-b92e-47891c2ff281-operator-scripts\") pod \"nova-cell1-61a4-account-create-update-rj4bg\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.851562 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62qxs\" (UniqueName: \"kubernetes.io/projected/179577e4-e654-46ef-b92e-47891c2ff281-kube-api-access-62qxs\") pod \"nova-cell1-61a4-account-create-update-rj4bg\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940328 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-internal-tls-certs\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940371 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-combined-ca-bundle\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940437 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-logs\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940469 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-config-data\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940545 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlfsl\" (UniqueName: \"kubernetes.io/projected/c15d3213-8b81-4da3-851b-a6927b51f7aa-kube-api-access-xlfsl\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940664 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-scripts\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940697 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.940726 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-httpd-run\") pod \"c15d3213-8b81-4da3-851b-a6927b51f7aa\" (UID: \"c15d3213-8b81-4da3-851b-a6927b51f7aa\") " Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.941663 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.942859 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-logs" (OuterVolumeSpecName: "logs") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.945674 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15d3213-8b81-4da3-851b-a6927b51f7aa-kube-api-access-xlfsl" (OuterVolumeSpecName: "kube-api-access-xlfsl") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "kube-api-access-xlfsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.951088 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-scripts" (OuterVolumeSpecName: "scripts") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.956402 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:52 crc kubenswrapper[4675]: I0219 09:02:52.973730 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.001213 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.045346 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlfsl\" (UniqueName: \"kubernetes.io/projected/c15d3213-8b81-4da3-851b-a6927b51f7aa-kube-api-access-xlfsl\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.045377 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.045411 4675 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.045421 4675 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.045434 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.045445 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c15d3213-8b81-4da3-851b-a6927b51f7aa-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.059286 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.092291 4675 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.122144 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-config-data" (OuterVolumeSpecName: "config-data") pod "c15d3213-8b81-4da3-851b-a6927b51f7aa" (UID: "c15d3213-8b81-4da3-851b-a6927b51f7aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.139317 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79aee500-9d06-41e6-8578-d92997ffbf71" path="/var/lib/kubelet/pods/79aee500-9d06-41e6-8578-d92997ffbf71/volumes" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.146748 4675 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.146781 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15d3213-8b81-4da3-851b-a6927b51f7aa-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.146791 4675 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.216133 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vk5pq"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.298523 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e4fa-account-create-update-rbg2p"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.476778 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c15d3213-8b81-4da3-851b-a6927b51f7aa","Type":"ContainerDied","Data":"bc148688259049807e14c163f7d8fbc3891c7e1859418ea8ff56bb47a0c57596"} Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.476851 4675 scope.go:117] "RemoveContainer" containerID="4491d6009caa431167f30a62aef780ddc463bf8377c2eda805101714b027a37a" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.477074 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: W0219 09:02:53.478402 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod461c89f2_a273_4964_b6fa_e6a858c7b97c.slice/crio-22c1ca0032f2ce15f6a5a0c365b5b9a6197abdf79e23e7909d14b56c1164127d WatchSource:0}: Error finding container 22c1ca0032f2ce15f6a5a0c365b5b9a6197abdf79e23e7909d14b56c1164127d: Status 404 returned error can't find the container with id 22c1ca0032f2ce15f6a5a0c365b5b9a6197abdf79e23e7909d14b56c1164127d Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.484924 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-q2tv7"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.488969 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerStarted","Data":"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f"} Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.490404 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vk5pq" event={"ID":"32253b6f-fc03-402c-bd55-8d49e4e3a7f1","Type":"ContainerStarted","Data":"427da18838032ead29525f212346b4c5fba9adbb28469ace231a8888da38b1a8"} Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.495990 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" event={"ID":"8b190a6d-42fd-487a-a90d-423a63e23603","Type":"ContainerStarted","Data":"630a906f144a1984c9c56696bb2216ae6735b01678c48cf838a33e6d99fdf98f"} Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.504453 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-crs9q"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.591064 4675 scope.go:117] "RemoveContainer" containerID="c966f779a7de37b0c7a32b6548053926814c76c76852307d77a6619f202d603a" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.602693 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.622492 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.645624 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:53 crc kubenswrapper[4675]: E0219 09:02:53.650164 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-log" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.650205 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-log" Feb 19 09:02:53 crc kubenswrapper[4675]: E0219 09:02:53.650248 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-httpd" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.650256 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-httpd" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.650467 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-log" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.650498 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" containerName="glance-httpd" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.651614 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.656251 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0a16-account-create-update-n7pz4"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.659110 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.659398 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.665143 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.743381 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-61a4-account-create-update-rj4bg"] Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.758862 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.759118 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a2c1f8-322b-40e7-a577-ed775f9a1565-logs\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.759201 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.759296 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.759425 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.759516 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.759725 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a2c1f8-322b-40e7-a577-ed775f9a1565-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.759844 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d86xm\" (UniqueName: \"kubernetes.io/projected/91a2c1f8-322b-40e7-a577-ed775f9a1565-kube-api-access-d86xm\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: W0219 09:02:53.780618 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod179577e4_e654_46ef_b92e_47891c2ff281.slice/crio-75f104dd35ad72b656471aca4d4ec15e64b2bcd890e4f0b9f28f6ed650e2eb0e WatchSource:0}: Error finding container 75f104dd35ad72b656471aca4d4ec15e64b2bcd890e4f0b9f28f6ed650e2eb0e: Status 404 returned error can't find the container with id 75f104dd35ad72b656471aca4d4ec15e64b2bcd890e4f0b9f28f6ed650e2eb0e Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.861205 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d86xm\" (UniqueName: \"kubernetes.io/projected/91a2c1f8-322b-40e7-a577-ed775f9a1565-kube-api-access-d86xm\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.861888 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.861951 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a2c1f8-322b-40e7-a577-ed775f9a1565-logs\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.861974 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.861993 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.862028 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.862048 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.862915 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.863031 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91a2c1f8-322b-40e7-a577-ed775f9a1565-logs\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.863860 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a2c1f8-322b-40e7-a577-ed775f9a1565-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.864799 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/91a2c1f8-322b-40e7-a577-ed775f9a1565-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.894026 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.896183 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.897793 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-scripts\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.913460 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a2c1f8-322b-40e7-a577-ed775f9a1565-config-data\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.920728 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d86xm\" (UniqueName: \"kubernetes.io/projected/91a2c1f8-322b-40e7-a577-ed775f9a1565-kube-api-access-d86xm\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:53 crc kubenswrapper[4675]: I0219 09:02:53.978411 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"91a2c1f8-322b-40e7-a577-ed775f9a1565\") " pod="openstack/glance-default-internal-api-0" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.044333 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-57f4c8679f-grtkm" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.116689 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.161146 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-565d4bc8d6-fvnsk"] Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.164015 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-565d4bc8d6-fvnsk" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-api" containerID="cri-o://681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571" gracePeriod=30 Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.164119 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-565d4bc8d6-fvnsk" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-httpd" containerID="cri-o://0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a" gracePeriod=30 Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.522995 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" event={"ID":"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e","Type":"ContainerStarted","Data":"ca09d3bbd549d2bc851a6704d9adb6c57182f2f3c86ff524a2f2887c95e77a57"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.523361 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" event={"ID":"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e","Type":"ContainerStarted","Data":"f4c72d39a7fef52e640ff199916f274f347a8fbc392b048e216ab02ff2c9444a"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.543060 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerStarted","Data":"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.548381 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" event={"ID":"179577e4-e654-46ef-b92e-47891c2ff281","Type":"ContainerStarted","Data":"2bfb00452299892a3c507150c5a2f51c41e62e03a71e43568f7b777267fe6ba0"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.548428 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" event={"ID":"179577e4-e654-46ef-b92e-47891c2ff281","Type":"ContainerStarted","Data":"75f104dd35ad72b656471aca4d4ec15e64b2bcd890e4f0b9f28f6ed650e2eb0e"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.560993 4675 generic.go:334] "Generic (PLEG): container finished" podID="32253b6f-fc03-402c-bd55-8d49e4e3a7f1" containerID="e9ff2ea7430ecd53359a998359ec420e76d30023669c91d97a57a414a2354b1b" exitCode=0 Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.561051 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vk5pq" event={"ID":"32253b6f-fc03-402c-bd55-8d49e4e3a7f1","Type":"ContainerDied","Data":"e9ff2ea7430ecd53359a998359ec420e76d30023669c91d97a57a414a2354b1b"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.562479 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" podStartSLOduration=2.5624539090000003 podStartE2EDuration="2.562453909s" podCreationTimestamp="2026-02-19 09:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:54.543243453 +0000 UTC m=+1156.170333721" watchObservedRunningTime="2026-02-19 09:02:54.562453909 +0000 UTC m=+1156.189544177" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.566430 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" event={"ID":"8b190a6d-42fd-487a-a90d-423a63e23603","Type":"ContainerStarted","Data":"9a5212464f4459a1bf7820a513a2f4e542f9da83989bf7a03af93f45c8254467"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.572632 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-q2tv7" event={"ID":"461c89f2-a273-4964-b6fa-e6a858c7b97c","Type":"ContainerStarted","Data":"22c920764004a79203fd0bfb3c5c316f352faa8b894cab5a2e245f6c6ab95eac"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.572698 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-q2tv7" event={"ID":"461c89f2-a273-4964-b6fa-e6a858c7b97c","Type":"ContainerStarted","Data":"22c1ca0032f2ce15f6a5a0c365b5b9a6197abdf79e23e7909d14b56c1164127d"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.602041 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" podStartSLOduration=2.601906006 podStartE2EDuration="2.601906006s" podCreationTimestamp="2026-02-19 09:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:54.582484926 +0000 UTC m=+1156.209575194" watchObservedRunningTime="2026-02-19 09:02:54.601906006 +0000 UTC m=+1156.228996274" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.611769 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-crs9q" event={"ID":"de9a8c34-ceea-4aac-91ca-ff7a8064c73b","Type":"ContainerStarted","Data":"ce44f8999c65da848864ac9d9109736b9e5c4a906ed6d0f7c95e31f6f092dc1a"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.611815 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-crs9q" event={"ID":"de9a8c34-ceea-4aac-91ca-ff7a8064c73b","Type":"ContainerStarted","Data":"d1b1bc7bf6a4f9b953dfd9d0976910b287c2325116d434f31e1e4a64e708226a"} Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.613931 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-q2tv7" podStartSLOduration=2.613890228 podStartE2EDuration="2.613890228s" podCreationTimestamp="2026-02-19 09:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:54.600023266 +0000 UTC m=+1156.227113534" watchObservedRunningTime="2026-02-19 09:02:54.613890228 +0000 UTC m=+1156.240980496" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.621185 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" podStartSLOduration=2.621168023 podStartE2EDuration="2.621168023s" podCreationTimestamp="2026-02-19 09:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:54.617211957 +0000 UTC m=+1156.244302225" watchObservedRunningTime="2026-02-19 09:02:54.621168023 +0000 UTC m=+1156.248258281" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.680724 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-crs9q" podStartSLOduration=3.680699229 podStartE2EDuration="3.680699229s" podCreationTimestamp="2026-02-19 09:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:54.668995446 +0000 UTC m=+1156.296085714" watchObservedRunningTime="2026-02-19 09:02:54.680699229 +0000 UTC m=+1156.307789497" Feb 19 09:02:54 crc kubenswrapper[4675]: I0219 09:02:54.838895 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 19 09:02:54 crc kubenswrapper[4675]: W0219 09:02:54.894160 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91a2c1f8_322b_40e7_a577_ed775f9a1565.slice/crio-83a976ef1ae1557bc0dae91cd986accaa1aa07c926ff7c47abde11111bae3d5a WatchSource:0}: Error finding container 83a976ef1ae1557bc0dae91cd986accaa1aa07c926ff7c47abde11111bae3d5a: Status 404 returned error can't find the container with id 83a976ef1ae1557bc0dae91cd986accaa1aa07c926ff7c47abde11111bae3d5a Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.115759 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c15d3213-8b81-4da3-851b-a6927b51f7aa" path="/var/lib/kubelet/pods/c15d3213-8b81-4da3-851b-a6927b51f7aa/volumes" Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.629809 4675 generic.go:334] "Generic (PLEG): container finished" podID="179577e4-e654-46ef-b92e-47891c2ff281" containerID="2bfb00452299892a3c507150c5a2f51c41e62e03a71e43568f7b777267fe6ba0" exitCode=0 Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.630053 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" event={"ID":"179577e4-e654-46ef-b92e-47891c2ff281","Type":"ContainerDied","Data":"2bfb00452299892a3c507150c5a2f51c41e62e03a71e43568f7b777267fe6ba0"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.633737 4675 generic.go:334] "Generic (PLEG): container finished" podID="8b190a6d-42fd-487a-a90d-423a63e23603" containerID="9a5212464f4459a1bf7820a513a2f4e542f9da83989bf7a03af93f45c8254467" exitCode=0 Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.633865 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" event={"ID":"8b190a6d-42fd-487a-a90d-423a63e23603","Type":"ContainerDied","Data":"9a5212464f4459a1bf7820a513a2f4e542f9da83989bf7a03af93f45c8254467"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.639002 4675 generic.go:334] "Generic (PLEG): container finished" podID="461c89f2-a273-4964-b6fa-e6a858c7b97c" containerID="22c920764004a79203fd0bfb3c5c316f352faa8b894cab5a2e245f6c6ab95eac" exitCode=0 Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.639049 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-q2tv7" event={"ID":"461c89f2-a273-4964-b6fa-e6a858c7b97c","Type":"ContainerDied","Data":"22c920764004a79203fd0bfb3c5c316f352faa8b894cab5a2e245f6c6ab95eac"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.642257 4675 generic.go:334] "Generic (PLEG): container finished" podID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerID="0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a" exitCode=0 Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.642319 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-565d4bc8d6-fvnsk" event={"ID":"5bfdd470-3f89-435e-9781-2543a8d69a5b","Type":"ContainerDied","Data":"0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.644034 4675 generic.go:334] "Generic (PLEG): container finished" podID="de9a8c34-ceea-4aac-91ca-ff7a8064c73b" containerID="ce44f8999c65da848864ac9d9109736b9e5c4a906ed6d0f7c95e31f6f092dc1a" exitCode=0 Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.644060 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-crs9q" event={"ID":"de9a8c34-ceea-4aac-91ca-ff7a8064c73b","Type":"ContainerDied","Data":"ce44f8999c65da848864ac9d9109736b9e5c4a906ed6d0f7c95e31f6f092dc1a"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.645477 4675 generic.go:334] "Generic (PLEG): container finished" podID="57dedd62-6c43-4bc5-a2f8-e98d20a8e73e" containerID="ca09d3bbd549d2bc851a6704d9adb6c57182f2f3c86ff524a2f2887c95e77a57" exitCode=0 Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.645521 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" event={"ID":"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e","Type":"ContainerDied","Data":"ca09d3bbd549d2bc851a6704d9adb6c57182f2f3c86ff524a2f2887c95e77a57"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.647989 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerStarted","Data":"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.652858 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91a2c1f8-322b-40e7-a577-ed775f9a1565","Type":"ContainerStarted","Data":"96b520034e8871ecf647938de300d9d14a7b82b687e5954197eba9d7936d823d"} Feb 19 09:02:55 crc kubenswrapper[4675]: I0219 09:02:55.652927 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91a2c1f8-322b-40e7-a577-ed775f9a1565","Type":"ContainerStarted","Data":"83a976ef1ae1557bc0dae91cd986accaa1aa07c926ff7c47abde11111bae3d5a"} Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.147786 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.153002 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc9wh\" (UniqueName: \"kubernetes.io/projected/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-kube-api-access-dc9wh\") pod \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.153065 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-operator-scripts\") pod \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\" (UID: \"32253b6f-fc03-402c-bd55-8d49e4e3a7f1\") " Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.156656 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32253b6f-fc03-402c-bd55-8d49e4e3a7f1" (UID: "32253b6f-fc03-402c-bd55-8d49e4e3a7f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.160887 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-kube-api-access-dc9wh" (OuterVolumeSpecName: "kube-api-access-dc9wh") pod "32253b6f-fc03-402c-bd55-8d49e4e3a7f1" (UID: "32253b6f-fc03-402c-bd55-8d49e4e3a7f1"). InnerVolumeSpecName "kube-api-access-dc9wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.263826 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc9wh\" (UniqueName: \"kubernetes.io/projected/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-kube-api-access-dc9wh\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.263865 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32253b6f-fc03-402c-bd55-8d49e4e3a7f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.664190 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vk5pq" Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.664204 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vk5pq" event={"ID":"32253b6f-fc03-402c-bd55-8d49e4e3a7f1","Type":"ContainerDied","Data":"427da18838032ead29525f212346b4c5fba9adbb28469ace231a8888da38b1a8"} Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.666272 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="427da18838032ead29525f212346b4c5fba9adbb28469ace231a8888da38b1a8" Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.669106 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"91a2c1f8-322b-40e7-a577-ed775f9a1565","Type":"ContainerStarted","Data":"b5fab67015e588c322a26054f996429bb188dc135e4687489c740ffc282075df"} Feb 19 09:02:56 crc kubenswrapper[4675]: I0219 09:02:56.707327 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.707305376 podStartE2EDuration="3.707305376s" podCreationTimestamp="2026-02-19 09:02:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:02:56.697226115 +0000 UTC m=+1158.324316383" watchObservedRunningTime="2026-02-19 09:02:56.707305376 +0000 UTC m=+1158.334395644" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.137243 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.227062 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.169:8776/healthcheck\": dial tcp 10.217.0.169:8776: connect: connection refused" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.287881 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62qxs\" (UniqueName: \"kubernetes.io/projected/179577e4-e654-46ef-b92e-47891c2ff281-kube-api-access-62qxs\") pod \"179577e4-e654-46ef-b92e-47891c2ff281\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.288306 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/179577e4-e654-46ef-b92e-47891c2ff281-operator-scripts\") pod \"179577e4-e654-46ef-b92e-47891c2ff281\" (UID: \"179577e4-e654-46ef-b92e-47891c2ff281\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.289763 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/179577e4-e654-46ef-b92e-47891c2ff281-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "179577e4-e654-46ef-b92e-47891c2ff281" (UID: "179577e4-e654-46ef-b92e-47891c2ff281"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.297867 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/179577e4-e654-46ef-b92e-47891c2ff281-kube-api-access-62qxs" (OuterVolumeSpecName: "kube-api-access-62qxs") pod "179577e4-e654-46ef-b92e-47891c2ff281" (UID: "179577e4-e654-46ef-b92e-47891c2ff281"). InnerVolumeSpecName "kube-api-access-62qxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.301217 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.345432 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.391147 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/179577e4-e654-46ef-b92e-47891c2ff281-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.391183 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62qxs\" (UniqueName: \"kubernetes.io/projected/179577e4-e654-46ef-b92e-47891c2ff281-kube-api-access-62qxs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.394340 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.412775 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.492312 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/461c89f2-a273-4964-b6fa-e6a858c7b97c-operator-scripts\") pod \"461c89f2-a273-4964-b6fa-e6a858c7b97c\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.492390 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-operator-scripts\") pod \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.492506 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxhtn\" (UniqueName: \"kubernetes.io/projected/461c89f2-a273-4964-b6fa-e6a858c7b97c-kube-api-access-vxhtn\") pod \"461c89f2-a273-4964-b6fa-e6a858c7b97c\" (UID: \"461c89f2-a273-4964-b6fa-e6a858c7b97c\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.492603 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sbbg\" (UniqueName: \"kubernetes.io/projected/8b190a6d-42fd-487a-a90d-423a63e23603-kube-api-access-2sbbg\") pod \"8b190a6d-42fd-487a-a90d-423a63e23603\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.492623 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b190a6d-42fd-487a-a90d-423a63e23603-operator-scripts\") pod \"8b190a6d-42fd-487a-a90d-423a63e23603\" (UID: \"8b190a6d-42fd-487a-a90d-423a63e23603\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.492658 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxc9g\" (UniqueName: \"kubernetes.io/projected/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-kube-api-access-gxc9g\") pod \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\" (UID: \"de9a8c34-ceea-4aac-91ca-ff7a8064c73b\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.493115 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b190a6d-42fd-487a-a90d-423a63e23603-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b190a6d-42fd-487a-a90d-423a63e23603" (UID: "8b190a6d-42fd-487a-a90d-423a63e23603"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.493337 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/461c89f2-a273-4964-b6fa-e6a858c7b97c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "461c89f2-a273-4964-b6fa-e6a858c7b97c" (UID: "461c89f2-a273-4964-b6fa-e6a858c7b97c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.498058 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de9a8c34-ceea-4aac-91ca-ff7a8064c73b" (UID: "de9a8c34-ceea-4aac-91ca-ff7a8064c73b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.502197 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/461c89f2-a273-4964-b6fa-e6a858c7b97c-kube-api-access-vxhtn" (OuterVolumeSpecName: "kube-api-access-vxhtn") pod "461c89f2-a273-4964-b6fa-e6a858c7b97c" (UID: "461c89f2-a273-4964-b6fa-e6a858c7b97c"). InnerVolumeSpecName "kube-api-access-vxhtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.502273 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-kube-api-access-gxc9g" (OuterVolumeSpecName: "kube-api-access-gxc9g") pod "de9a8c34-ceea-4aac-91ca-ff7a8064c73b" (UID: "de9a8c34-ceea-4aac-91ca-ff7a8064c73b"). InnerVolumeSpecName "kube-api-access-gxc9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.505889 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b190a6d-42fd-487a-a90d-423a63e23603-kube-api-access-2sbbg" (OuterVolumeSpecName: "kube-api-access-2sbbg") pod "8b190a6d-42fd-487a-a90d-423a63e23603" (UID: "8b190a6d-42fd-487a-a90d-423a63e23603"). InnerVolumeSpecName "kube-api-access-2sbbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.554759 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.594829 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsp42\" (UniqueName: \"kubernetes.io/projected/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-kube-api-access-bsp42\") pod \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.594961 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-operator-scripts\") pod \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\" (UID: \"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.595739 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxhtn\" (UniqueName: \"kubernetes.io/projected/461c89f2-a273-4964-b6fa-e6a858c7b97c-kube-api-access-vxhtn\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.595765 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sbbg\" (UniqueName: \"kubernetes.io/projected/8b190a6d-42fd-487a-a90d-423a63e23603-kube-api-access-2sbbg\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.595781 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b190a6d-42fd-487a-a90d-423a63e23603-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.595793 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxc9g\" (UniqueName: \"kubernetes.io/projected/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-kube-api-access-gxc9g\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.595805 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/461c89f2-a273-4964-b6fa-e6a858c7b97c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.595816 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9a8c34-ceea-4aac-91ca-ff7a8064c73b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.596231 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "57dedd62-6c43-4bc5-a2f8-e98d20a8e73e" (UID: "57dedd62-6c43-4bc5-a2f8-e98d20a8e73e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.599303 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-kube-api-access-bsp42" (OuterVolumeSpecName: "kube-api-access-bsp42") pod "57dedd62-6c43-4bc5-a2f8-e98d20a8e73e" (UID: "57dedd62-6c43-4bc5-a2f8-e98d20a8e73e"). InnerVolumeSpecName "kube-api-access-bsp42". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.681705 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-crs9q" event={"ID":"de9a8c34-ceea-4aac-91ca-ff7a8064c73b","Type":"ContainerDied","Data":"d1b1bc7bf6a4f9b953dfd9d0976910b287c2325116d434f31e1e4a64e708226a"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.681770 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1b1bc7bf6a4f9b953dfd9d0976910b287c2325116d434f31e1e4a64e708226a" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.681830 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-crs9q" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.690766 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" event={"ID":"57dedd62-6c43-4bc5-a2f8-e98d20a8e73e","Type":"ContainerDied","Data":"f4c72d39a7fef52e640ff199916f274f347a8fbc392b048e216ab02ff2c9444a"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.691429 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4c72d39a7fef52e640ff199916f274f347a8fbc392b048e216ab02ff2c9444a" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.691680 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0a16-account-create-update-n7pz4" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697074 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwsvk\" (UniqueName: \"kubernetes.io/projected/457624d6-3a3d-406f-bc5e-626f1a0244ac-kube-api-access-nwsvk\") pod \"457624d6-3a3d-406f-bc5e-626f1a0244ac\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697120 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data\") pod \"457624d6-3a3d-406f-bc5e-626f1a0244ac\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697172 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-scripts\") pod \"457624d6-3a3d-406f-bc5e-626f1a0244ac\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697202 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-combined-ca-bundle\") pod \"457624d6-3a3d-406f-bc5e-626f1a0244ac\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697255 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/457624d6-3a3d-406f-bc5e-626f1a0244ac-logs\") pod \"457624d6-3a3d-406f-bc5e-626f1a0244ac\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697300 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data-custom\") pod \"457624d6-3a3d-406f-bc5e-626f1a0244ac\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697356 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/457624d6-3a3d-406f-bc5e-626f1a0244ac-etc-machine-id\") pod \"457624d6-3a3d-406f-bc5e-626f1a0244ac\" (UID: \"457624d6-3a3d-406f-bc5e-626f1a0244ac\") " Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697869 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsp42\" (UniqueName: \"kubernetes.io/projected/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-kube-api-access-bsp42\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697886 4675 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.697932 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457624d6-3a3d-406f-bc5e-626f1a0244ac-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "457624d6-3a3d-406f-bc5e-626f1a0244ac" (UID: "457624d6-3a3d-406f-bc5e-626f1a0244ac"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.701785 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/457624d6-3a3d-406f-bc5e-626f1a0244ac-logs" (OuterVolumeSpecName: "logs") pod "457624d6-3a3d-406f-bc5e-626f1a0244ac" (UID: "457624d6-3a3d-406f-bc5e-626f1a0244ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.702760 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/457624d6-3a3d-406f-bc5e-626f1a0244ac-kube-api-access-nwsvk" (OuterVolumeSpecName: "kube-api-access-nwsvk") pod "457624d6-3a3d-406f-bc5e-626f1a0244ac" (UID: "457624d6-3a3d-406f-bc5e-626f1a0244ac"). InnerVolumeSpecName "kube-api-access-nwsvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.712857 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "457624d6-3a3d-406f-bc5e-626f1a0244ac" (UID: "457624d6-3a3d-406f-bc5e-626f1a0244ac"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.712870 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-scripts" (OuterVolumeSpecName: "scripts") pod "457624d6-3a3d-406f-bc5e-626f1a0244ac" (UID: "457624d6-3a3d-406f-bc5e-626f1a0244ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.732242 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerStarted","Data":"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.732841 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.739703 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.739878 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-61a4-account-create-update-rj4bg" event={"ID":"179577e4-e654-46ef-b92e-47891c2ff281","Type":"ContainerDied","Data":"75f104dd35ad72b656471aca4d4ec15e64b2bcd890e4f0b9f28f6ed650e2eb0e"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.740115 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75f104dd35ad72b656471aca4d4ec15e64b2bcd890e4f0b9f28f6ed650e2eb0e" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.741847 4675 generic.go:334] "Generic (PLEG): container finished" podID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerID="3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136" exitCode=137 Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.742050 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"457624d6-3a3d-406f-bc5e-626f1a0244ac","Type":"ContainerDied","Data":"3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.742098 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"457624d6-3a3d-406f-bc5e-626f1a0244ac","Type":"ContainerDied","Data":"b79b16cede22a55cbaa6ee0b64f899325cf976149e485661d3144dd0a755d270"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.742123 4675 scope.go:117] "RemoveContainer" containerID="3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.742557 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.747390 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" event={"ID":"8b190a6d-42fd-487a-a90d-423a63e23603","Type":"ContainerDied","Data":"630a906f144a1984c9c56696bb2216ae6735b01678c48cf838a33e6d99fdf98f"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.747445 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="630a906f144a1984c9c56696bb2216ae6735b01678c48cf838a33e6d99fdf98f" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.747568 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4fa-account-create-update-rbg2p" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.754036 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-q2tv7" event={"ID":"461c89f2-a273-4964-b6fa-e6a858c7b97c","Type":"ContainerDied","Data":"22c1ca0032f2ce15f6a5a0c365b5b9a6197abdf79e23e7909d14b56c1164127d"} Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.754098 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c1ca0032f2ce15f6a5a0c365b5b9a6197abdf79e23e7909d14b56c1164127d" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.765734 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "457624d6-3a3d-406f-bc5e-626f1a0244ac" (UID: "457624d6-3a3d-406f-bc5e-626f1a0244ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.771556 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-q2tv7" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.810119 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwsvk\" (UniqueName: \"kubernetes.io/projected/457624d6-3a3d-406f-bc5e-626f1a0244ac-kube-api-access-nwsvk\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.810778 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.810810 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.810824 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/457624d6-3a3d-406f-bc5e-626f1a0244ac-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.810852 4675 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.810869 4675 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/457624d6-3a3d-406f-bc5e-626f1a0244ac-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.832566 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.255083649 podStartE2EDuration="6.832536331s" podCreationTimestamp="2026-02-19 09:02:51 +0000 UTC" firstStartedPulling="2026-02-19 09:02:52.355272939 +0000 UTC m=+1153.982363207" lastFinishedPulling="2026-02-19 09:02:56.932725621 +0000 UTC m=+1158.559815889" observedRunningTime="2026-02-19 09:02:57.766905081 +0000 UTC m=+1159.393995349" watchObservedRunningTime="2026-02-19 09:02:57.832536331 +0000 UTC m=+1159.459626599" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.845120 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data" (OuterVolumeSpecName: "config-data") pod "457624d6-3a3d-406f-bc5e-626f1a0244ac" (UID: "457624d6-3a3d-406f-bc5e-626f1a0244ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.869895 4675 scope.go:117] "RemoveContainer" containerID="e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.912366 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/457624d6-3a3d-406f-bc5e-626f1a0244ac-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.914552 4675 scope.go:117] "RemoveContainer" containerID="3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136" Feb 19 09:02:57 crc kubenswrapper[4675]: E0219 09:02:57.915218 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136\": container with ID starting with 3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136 not found: ID does not exist" containerID="3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.915286 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136"} err="failed to get container status \"3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136\": rpc error: code = NotFound desc = could not find container \"3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136\": container with ID starting with 3e7a8376ea2a777a06ae227f215a853fb5f6c5bdc198514af4d6f320ddab1136 not found: ID does not exist" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.915328 4675 scope.go:117] "RemoveContainer" containerID="e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6" Feb 19 09:02:57 crc kubenswrapper[4675]: E0219 09:02:57.915720 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6\": container with ID starting with e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6 not found: ID does not exist" containerID="e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6" Feb 19 09:02:57 crc kubenswrapper[4675]: I0219 09:02:57.915753 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6"} err="failed to get container status \"e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6\": rpc error: code = NotFound desc = could not find container \"e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6\": container with ID starting with e08bdcd337ac85debcdbb8248bd18b0b35d4eb9f4d4beca82d244b5e496025b6 not found: ID does not exist" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.079482 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.098745 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118050 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118629 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32253b6f-fc03-402c-bd55-8d49e4e3a7f1" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118667 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="32253b6f-fc03-402c-bd55-8d49e4e3a7f1" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118683 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dedd62-6c43-4bc5-a2f8-e98d20a8e73e" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118691 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dedd62-6c43-4bc5-a2f8-e98d20a8e73e" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118703 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api-log" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118710 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api-log" Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118718 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461c89f2-a273-4964-b6fa-e6a858c7b97c" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118724 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="461c89f2-a273-4964-b6fa-e6a858c7b97c" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118740 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b190a6d-42fd-487a-a90d-423a63e23603" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118747 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b190a6d-42fd-487a-a90d-423a63e23603" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118767 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179577e4-e654-46ef-b92e-47891c2ff281" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118773 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="179577e4-e654-46ef-b92e-47891c2ff281" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118795 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118803 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api" Feb 19 09:02:58 crc kubenswrapper[4675]: E0219 09:02:58.118816 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9a8c34-ceea-4aac-91ca-ff7a8064c73b" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.118822 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9a8c34-ceea-4aac-91ca-ff7a8064c73b" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119017 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api-log" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119038 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="179577e4-e654-46ef-b92e-47891c2ff281" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119049 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="461c89f2-a273-4964-b6fa-e6a858c7b97c" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119061 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b190a6d-42fd-487a-a90d-423a63e23603" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119086 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9a8c34-ceea-4aac-91ca-ff7a8064c73b" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119101 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="32253b6f-fc03-402c-bd55-8d49e4e3a7f1" containerName="mariadb-database-create" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119115 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dedd62-6c43-4bc5-a2f8-e98d20a8e73e" containerName="mariadb-account-create-update" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.119127 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" containerName="cinder-api" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.120327 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.122534 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.122863 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.122999 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.145175 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.216944 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.217001 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-logs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.217926 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-scripts\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.217956 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-config-data\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.218021 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.218195 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.218235 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-config-data-custom\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.218306 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.218336 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djswp\" (UniqueName: \"kubernetes.io/projected/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-kube-api-access-djswp\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320051 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320541 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-config-data-custom\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320608 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320651 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djswp\" (UniqueName: \"kubernetes.io/projected/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-kube-api-access-djswp\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320689 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320748 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-logs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320807 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-scripts\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320826 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-config-data\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320859 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.320947 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.321711 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-logs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.325391 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.326239 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-scripts\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.326441 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-config-data-custom\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.327603 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.329117 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-config-data\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.333856 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.343442 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djswp\" (UniqueName: \"kubernetes.io/projected/3b83eea9-3ec2-44bd-8498-5c5604e0eac7-kube-api-access-djswp\") pod \"cinder-api-0\" (UID: \"3b83eea9-3ec2-44bd-8498-5c5604e0eac7\") " pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.442973 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 19 09:02:58 crc kubenswrapper[4675]: I0219 09:02:58.934798 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.088800 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.089374 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-log" containerID="cri-o://ec9226b305f0ce2cd0d2f932df6518705ee9d7c236b0bd4e5ad19b7f29b30918" gracePeriod=30 Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.089517 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-httpd" containerID="cri-o://c9173343cbea7deb828160fa7599a578a21600e26e0722f6d0dfaca422335b38" gracePeriod=30 Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.116685 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="457624d6-3a3d-406f-bc5e-626f1a0244ac" path="/var/lib/kubelet/pods/457624d6-3a3d-406f-bc5e-626f1a0244ac/volumes" Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.800199 4675 generic.go:334] "Generic (PLEG): container finished" podID="d2589e24-7bb6-49ac-a98c-55be590405df" containerID="ec9226b305f0ce2cd0d2f932df6518705ee9d7c236b0bd4e5ad19b7f29b30918" exitCode=143 Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.800273 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2589e24-7bb6-49ac-a98c-55be590405df","Type":"ContainerDied","Data":"ec9226b305f0ce2cd0d2f932df6518705ee9d7c236b0bd4e5ad19b7f29b30918"} Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.810918 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3b83eea9-3ec2-44bd-8498-5c5604e0eac7","Type":"ContainerStarted","Data":"634e905afe8b35d61a1615c963fe7754b9b6a419a9d76d6d77b82373fe1af912"} Feb 19 09:02:59 crc kubenswrapper[4675]: I0219 09:02:59.810965 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3b83eea9-3ec2-44bd-8498-5c5604e0eac7","Type":"ContainerStarted","Data":"775c42a57cb741ad25fc79a270cf9c5eaef899783c849804403b1b46d26e6c49"} Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.057739 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bbbd6c858-v6gv9" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.574630 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.727371 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-httpd-config\") pod \"5bfdd470-3f89-435e-9781-2543a8d69a5b\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.727505 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-ovndb-tls-certs\") pod \"5bfdd470-3f89-435e-9781-2543a8d69a5b\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.727699 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-config\") pod \"5bfdd470-3f89-435e-9781-2543a8d69a5b\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.727752 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-combined-ca-bundle\") pod \"5bfdd470-3f89-435e-9781-2543a8d69a5b\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.727807 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljfcf\" (UniqueName: \"kubernetes.io/projected/5bfdd470-3f89-435e-9781-2543a8d69a5b-kube-api-access-ljfcf\") pod \"5bfdd470-3f89-435e-9781-2543a8d69a5b\" (UID: \"5bfdd470-3f89-435e-9781-2543a8d69a5b\") " Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.735863 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5bfdd470-3f89-435e-9781-2543a8d69a5b" (UID: "5bfdd470-3f89-435e-9781-2543a8d69a5b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.763889 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bfdd470-3f89-435e-9781-2543a8d69a5b-kube-api-access-ljfcf" (OuterVolumeSpecName: "kube-api-access-ljfcf") pod "5bfdd470-3f89-435e-9781-2543a8d69a5b" (UID: "5bfdd470-3f89-435e-9781-2543a8d69a5b"). InnerVolumeSpecName "kube-api-access-ljfcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.830129 4675 generic.go:334] "Generic (PLEG): container finished" podID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerID="681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571" exitCode=0 Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.830269 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-565d4bc8d6-fvnsk" event={"ID":"5bfdd470-3f89-435e-9781-2543a8d69a5b","Type":"ContainerDied","Data":"681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571"} Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.830304 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-565d4bc8d6-fvnsk" event={"ID":"5bfdd470-3f89-435e-9781-2543a8d69a5b","Type":"ContainerDied","Data":"bd38504998f99efd6f6afc715e51da999b836e134a8db449e15e304a524307d2"} Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.830356 4675 scope.go:117] "RemoveContainer" containerID="0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.830600 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-565d4bc8d6-fvnsk" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.835155 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3b83eea9-3ec2-44bd-8498-5c5604e0eac7","Type":"ContainerStarted","Data":"9461f377716212349cdbd2ca9419730fef918a4f23812e4a3866c92d4d45699f"} Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.835520 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.835603 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljfcf\" (UniqueName: \"kubernetes.io/projected/5bfdd470-3f89-435e-9781-2543a8d69a5b-kube-api-access-ljfcf\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.836116 4675 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.845216 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-config" (OuterVolumeSpecName: "config") pod "5bfdd470-3f89-435e-9781-2543a8d69a5b" (UID: "5bfdd470-3f89-435e-9781-2543a8d69a5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.878741 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.878720768 podStartE2EDuration="2.878720768s" podCreationTimestamp="2026-02-19 09:02:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:00.857830568 +0000 UTC m=+1162.484920836" watchObservedRunningTime="2026-02-19 09:03:00.878720768 +0000 UTC m=+1162.505811036" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.888443 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5bfdd470-3f89-435e-9781-2543a8d69a5b" (UID: "5bfdd470-3f89-435e-9781-2543a8d69a5b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.889216 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bfdd470-3f89-435e-9781-2543a8d69a5b" (UID: "5bfdd470-3f89-435e-9781-2543a8d69a5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.940020 4675 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.940049 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:00 crc kubenswrapper[4675]: I0219 09:03:00.940060 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfdd470-3f89-435e-9781-2543a8d69a5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:01 crc kubenswrapper[4675]: I0219 09:03:01.007893 4675 scope.go:117] "RemoveContainer" containerID="681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571" Feb 19 09:03:01 crc kubenswrapper[4675]: I0219 09:03:01.041422 4675 scope.go:117] "RemoveContainer" containerID="0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a" Feb 19 09:03:01 crc kubenswrapper[4675]: E0219 09:03:01.042009 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a\": container with ID starting with 0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a not found: ID does not exist" containerID="0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a" Feb 19 09:03:01 crc kubenswrapper[4675]: I0219 09:03:01.042106 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a"} err="failed to get container status \"0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a\": rpc error: code = NotFound desc = could not find container \"0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a\": container with ID starting with 0dbe5995fd05383e89401c8d471836118abf247df49f95c09dc3343609977c1a not found: ID does not exist" Feb 19 09:03:01 crc kubenswrapper[4675]: I0219 09:03:01.042155 4675 scope.go:117] "RemoveContainer" containerID="681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571" Feb 19 09:03:01 crc kubenswrapper[4675]: E0219 09:03:01.042484 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571\": container with ID starting with 681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571 not found: ID does not exist" containerID="681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571" Feb 19 09:03:01 crc kubenswrapper[4675]: I0219 09:03:01.042507 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571"} err="failed to get container status \"681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571\": rpc error: code = NotFound desc = could not find container \"681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571\": container with ID starting with 681f214101781222dabf6dcbf3fb6f1a7e1f694836f0cc5c0304b7da56ea9571 not found: ID does not exist" Feb 19 09:03:01 crc kubenswrapper[4675]: I0219 09:03:01.165864 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-565d4bc8d6-fvnsk"] Feb 19 09:03:01 crc kubenswrapper[4675]: I0219 09:03:01.180179 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-565d4bc8d6-fvnsk"] Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.240172 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9292/healthcheck\": read tcp 10.217.0.2:35800->10.217.0.158:9292: read: connection reset by peer" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.240189 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.158:9292/healthcheck\": read tcp 10.217.0.2:35802->10.217.0.158:9292: read: connection reset by peer" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.828125 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dlxqd"] Feb 19 09:03:02 crc kubenswrapper[4675]: E0219 09:03:02.829656 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-api" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.829785 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-api" Feb 19 09:03:02 crc kubenswrapper[4675]: E0219 09:03:02.829859 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-httpd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.829923 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-httpd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.830183 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-api" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.830262 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" containerName="neutron-httpd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.830949 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.834095 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.835955 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.836148 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fgmtk" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.847623 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dlxqd"] Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.869451 4675 generic.go:334] "Generic (PLEG): container finished" podID="d2589e24-7bb6-49ac-a98c-55be590405df" containerID="c9173343cbea7deb828160fa7599a578a21600e26e0722f6d0dfaca422335b38" exitCode=0 Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.869493 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2589e24-7bb6-49ac-a98c-55be590405df","Type":"ContainerDied","Data":"c9173343cbea7deb828160fa7599a578a21600e26e0722f6d0dfaca422335b38"} Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.869520 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2589e24-7bb6-49ac-a98c-55be590405df","Type":"ContainerDied","Data":"eb01db82e678b1693be4bb1c8fafc358cb1aa1729e5fc4bd23c784f1b61c6bdb"} Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.869534 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb01db82e678b1693be4bb1c8fafc358cb1aa1729e5fc4bd23c784f1b61c6bdb" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.869664 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.878412 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j9vv\" (UniqueName: \"kubernetes.io/projected/f6c638bb-53d1-4c7a-889b-a506b89b202e-kube-api-access-2j9vv\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.878549 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-scripts\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.878679 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.878752 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-config-data\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.979915 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-logs\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.979983 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-public-tls-certs\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980030 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rztfj\" (UniqueName: \"kubernetes.io/projected/d2589e24-7bb6-49ac-a98c-55be590405df-kube-api-access-rztfj\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980092 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980192 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-combined-ca-bundle\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980207 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-config-data\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980253 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-httpd-run\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980314 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-scripts\") pod \"d2589e24-7bb6-49ac-a98c-55be590405df\" (UID: \"d2589e24-7bb6-49ac-a98c-55be590405df\") " Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980569 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j9vv\" (UniqueName: \"kubernetes.io/projected/f6c638bb-53d1-4c7a-889b-a506b89b202e-kube-api-access-2j9vv\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980904 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-scripts\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980951 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980595 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-logs" (OuterVolumeSpecName: "logs") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.980980 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-config-data\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.981196 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.981279 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.990936 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-scripts" (OuterVolumeSpecName: "scripts") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.991072 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2589e24-7bb6-49ac-a98c-55be590405df-kube-api-access-rztfj" (OuterVolumeSpecName: "kube-api-access-rztfj") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "kube-api-access-rztfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.991588 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-config-data\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.992588 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.994782 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 19 09:03:02 crc kubenswrapper[4675]: I0219 09:03:02.995218 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-scripts\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.014270 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j9vv\" (UniqueName: \"kubernetes.io/projected/f6c638bb-53d1-4c7a-889b-a506b89b202e-kube-api-access-2j9vv\") pod \"nova-cell0-conductor-db-sync-dlxqd\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.026635 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.052080 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-config-data" (OuterVolumeSpecName: "config-data") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.054742 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d2589e24-7bb6-49ac-a98c-55be590405df" (UID: "d2589e24-7bb6-49ac-a98c-55be590405df"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.083145 4675 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2589e24-7bb6-49ac-a98c-55be590405df-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.083181 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.083191 4675 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.083203 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rztfj\" (UniqueName: \"kubernetes.io/projected/d2589e24-7bb6-49ac-a98c-55be590405df-kube-api-access-rztfj\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.083240 4675 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.083249 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.083258 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2589e24-7bb6-49ac-a98c-55be590405df-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.110439 4675 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.114300 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bfdd470-3f89-435e-9781-2543a8d69a5b" path="/var/lib/kubelet/pods/5bfdd470-3f89-435e-9781-2543a8d69a5b/volumes" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.184821 4675 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.202184 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.674471 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dlxqd"] Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.878994 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.878963 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" event={"ID":"f6c638bb-53d1-4c7a-889b-a506b89b202e","Type":"ContainerStarted","Data":"905d2469db76fbd10e79379f73d1de371b3065bbe6677002fb5d3a67a559f247"} Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.911957 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.921747 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.960669 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:03:03 crc kubenswrapper[4675]: E0219 09:03:03.961084 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-log" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.961103 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-log" Feb 19 09:03:03 crc kubenswrapper[4675]: E0219 09:03:03.961135 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-httpd" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.961142 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-httpd" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.961318 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-httpd" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.961351 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" containerName="glance-log" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.962284 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.967659 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.968325 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 19 09:03:03 crc kubenswrapper[4675]: I0219 09:03:03.970753 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.000663 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.000878 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-config-data\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.000934 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.000956 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.001108 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjf2f\" (UniqueName: \"kubernetes.io/projected/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-kube-api-access-vjf2f\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.001179 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-logs\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.001228 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-scripts\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.001311 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.102661 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.103070 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-config-data\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.103086 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.103358 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.103415 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.103498 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjf2f\" (UniqueName: \"kubernetes.io/projected/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-kube-api-access-vjf2f\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.103556 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-logs\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.103597 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-scripts\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.104097 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-logs\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.104439 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.104839 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.110896 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-scripts\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.111576 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.112742 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-config-data\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.115989 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.116951 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.116989 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.125927 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjf2f\" (UniqueName: \"kubernetes.io/projected/0be92ca9-2a5c-4c33-b7e4-a440f157cd02-kube-api-access-vjf2f\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.145568 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0be92ca9-2a5c-4c33-b7e4-a440f157cd02\") " pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.170190 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.179325 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.289116 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.809296 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.811295 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-central-agent" containerID="cri-o://dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" gracePeriod=30 Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.811915 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="proxy-httpd" containerID="cri-o://641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" gracePeriod=30 Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.812072 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="sg-core" containerID="cri-o://24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" gracePeriod=30 Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.812255 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-notification-agent" containerID="cri-o://b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" gracePeriod=30 Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.887559 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.887617 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:04 crc kubenswrapper[4675]: I0219 09:03:04.905365 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 19 09:03:04 crc kubenswrapper[4675]: W0219 09:03:04.909775 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0be92ca9_2a5c_4c33_b7e4_a440f157cd02.slice/crio-d352b5b9e4e51449db1e0037137e7d177ef47a2535c89900fad03ef87da20f3c WatchSource:0}: Error finding container d352b5b9e4e51449db1e0037137e7d177ef47a2535c89900fad03ef87da20f3c: Status 404 returned error can't find the container with id d352b5b9e4e51449db1e0037137e7d177ef47a2535c89900fad03ef87da20f3c Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.115507 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2589e24-7bb6-49ac-a98c-55be590405df" path="/var/lib/kubelet/pods/d2589e24-7bb6-49ac-a98c-55be590405df/volumes" Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.921810 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923416 4675 generic.go:334] "Generic (PLEG): container finished" podID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerID="641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" exitCode=0 Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923450 4675 generic.go:334] "Generic (PLEG): container finished" podID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerID="24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" exitCode=2 Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923460 4675 generic.go:334] "Generic (PLEG): container finished" podID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerID="b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" exitCode=0 Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923471 4675 generic.go:334] "Generic (PLEG): container finished" podID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerID="dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" exitCode=0 Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923523 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerDied","Data":"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6"} Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923551 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerDied","Data":"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459"} Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923568 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerDied","Data":"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805"} Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923582 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerDied","Data":"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f"} Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923593 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5633bc1-203a-4dd9-9089-ee1e7a5a115e","Type":"ContainerDied","Data":"dea5c963c7d98038602e67b43efb3d753775dbd0ddf0aec511a29054c5ccfbe0"} Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.923610 4675 scope.go:117] "RemoveContainer" containerID="641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.930011 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0be92ca9-2a5c-4c33-b7e4-a440f157cd02","Type":"ContainerStarted","Data":"fa6159a1a0f8092ebda660a26920f77ee93b82705a9368cad9507cd7fc1401a0"} Feb 19 09:03:05 crc kubenswrapper[4675]: I0219 09:03:05.930059 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0be92ca9-2a5c-4c33-b7e4-a440f157cd02","Type":"ContainerStarted","Data":"d352b5b9e4e51449db1e0037137e7d177ef47a2535c89900fad03ef87da20f3c"} Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.002007 4675 scope.go:117] "RemoveContainer" containerID="24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.054913 4675 scope.go:117] "RemoveContainer" containerID="b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.055162 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-config-data\") pod \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.055333 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-log-httpd\") pod \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.055481 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-run-httpd\") pod \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.055742 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-combined-ca-bundle\") pod \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.055808 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-sg-core-conf-yaml\") pod \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.056208 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-scripts\") pod \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.056263 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk9dr\" (UniqueName: \"kubernetes.io/projected/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-kube-api-access-tk9dr\") pod \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\" (UID: \"d5633bc1-203a-4dd9-9089-ee1e7a5a115e\") " Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.056450 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d5633bc1-203a-4dd9-9089-ee1e7a5a115e" (UID: "d5633bc1-203a-4dd9-9089-ee1e7a5a115e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.056832 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d5633bc1-203a-4dd9-9089-ee1e7a5a115e" (UID: "d5633bc1-203a-4dd9-9089-ee1e7a5a115e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.058747 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.058770 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.063431 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-kube-api-access-tk9dr" (OuterVolumeSpecName: "kube-api-access-tk9dr") pod "d5633bc1-203a-4dd9-9089-ee1e7a5a115e" (UID: "d5633bc1-203a-4dd9-9089-ee1e7a5a115e"). InnerVolumeSpecName "kube-api-access-tk9dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.066278 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-scripts" (OuterVolumeSpecName: "scripts") pod "d5633bc1-203a-4dd9-9089-ee1e7a5a115e" (UID: "d5633bc1-203a-4dd9-9089-ee1e7a5a115e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.088823 4675 scope.go:117] "RemoveContainer" containerID="dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.101384 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d5633bc1-203a-4dd9-9089-ee1e7a5a115e" (UID: "d5633bc1-203a-4dd9-9089-ee1e7a5a115e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.141112 4675 scope.go:117] "RemoveContainer" containerID="641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" Feb 19 09:03:06 crc kubenswrapper[4675]: E0219 09:03:06.141622 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": container with ID starting with 641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6 not found: ID does not exist" containerID="641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.141669 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6"} err="failed to get container status \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": rpc error: code = NotFound desc = could not find container \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": container with ID starting with 641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.141689 4675 scope.go:117] "RemoveContainer" containerID="24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" Feb 19 09:03:06 crc kubenswrapper[4675]: E0219 09:03:06.141919 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": container with ID starting with 24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459 not found: ID does not exist" containerID="24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.141960 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459"} err="failed to get container status \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": rpc error: code = NotFound desc = could not find container \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": container with ID starting with 24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.141974 4675 scope.go:117] "RemoveContainer" containerID="b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" Feb 19 09:03:06 crc kubenswrapper[4675]: E0219 09:03:06.142170 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": container with ID starting with b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805 not found: ID does not exist" containerID="b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.142200 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805"} err="failed to get container status \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": rpc error: code = NotFound desc = could not find container \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": container with ID starting with b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.142216 4675 scope.go:117] "RemoveContainer" containerID="dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" Feb 19 09:03:06 crc kubenswrapper[4675]: E0219 09:03:06.142485 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": container with ID starting with dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f not found: ID does not exist" containerID="dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.142509 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f"} err="failed to get container status \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": rpc error: code = NotFound desc = could not find container \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": container with ID starting with dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.142527 4675 scope.go:117] "RemoveContainer" containerID="641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.142818 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6"} err="failed to get container status \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": rpc error: code = NotFound desc = could not find container \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": container with ID starting with 641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.142837 4675 scope.go:117] "RemoveContainer" containerID="24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.143043 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459"} err="failed to get container status \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": rpc error: code = NotFound desc = could not find container \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": container with ID starting with 24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.143061 4675 scope.go:117] "RemoveContainer" containerID="b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.143232 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805"} err="failed to get container status \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": rpc error: code = NotFound desc = could not find container \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": container with ID starting with b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.143249 4675 scope.go:117] "RemoveContainer" containerID="dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.143484 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f"} err="failed to get container status \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": rpc error: code = NotFound desc = could not find container \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": container with ID starting with dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.143501 4675 scope.go:117] "RemoveContainer" containerID="641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.144580 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6"} err="failed to get container status \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": rpc error: code = NotFound desc = could not find container \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": container with ID starting with 641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.144606 4675 scope.go:117] "RemoveContainer" containerID="24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.146857 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459"} err="failed to get container status \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": rpc error: code = NotFound desc = could not find container \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": container with ID starting with 24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.146895 4675 scope.go:117] "RemoveContainer" containerID="b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.147483 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805"} err="failed to get container status \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": rpc error: code = NotFound desc = could not find container \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": container with ID starting with b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.147502 4675 scope.go:117] "RemoveContainer" containerID="dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.147859 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f"} err="failed to get container status \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": rpc error: code = NotFound desc = could not find container \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": container with ID starting with dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.147879 4675 scope.go:117] "RemoveContainer" containerID="641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.149713 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6"} err="failed to get container status \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": rpc error: code = NotFound desc = could not find container \"641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6\": container with ID starting with 641935861e89a9562552f14283e26ad6fb73d48f60c5092c00c718c323fd74d6 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.149736 4675 scope.go:117] "RemoveContainer" containerID="24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.150324 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459"} err="failed to get container status \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": rpc error: code = NotFound desc = could not find container \"24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459\": container with ID starting with 24c01c952cc72e1b1828735ecd761f81a755d75e88acabef0d8201feec859459 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.150342 4675 scope.go:117] "RemoveContainer" containerID="b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.150796 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805"} err="failed to get container status \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": rpc error: code = NotFound desc = could not find container \"b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805\": container with ID starting with b4ee355d6a6f00a3cc05716dc221331515987e3098e31933bc05fd0a755b7805 not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.150817 4675 scope.go:117] "RemoveContainer" containerID="dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.151315 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f"} err="failed to get container status \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": rpc error: code = NotFound desc = could not find container \"dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f\": container with ID starting with dab2e2cc1a6279bbdc5e8b94f83e040407663e77b047bb82fdede86162ceef7f not found: ID does not exist" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.161616 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.161666 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.161677 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk9dr\" (UniqueName: \"kubernetes.io/projected/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-kube-api-access-tk9dr\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.183781 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5633bc1-203a-4dd9-9089-ee1e7a5a115e" (UID: "d5633bc1-203a-4dd9-9089-ee1e7a5a115e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.195504 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-config-data" (OuterVolumeSpecName: "config-data") pod "d5633bc1-203a-4dd9-9089-ee1e7a5a115e" (UID: "d5633bc1-203a-4dd9-9089-ee1e7a5a115e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.264363 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.264771 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5633bc1-203a-4dd9-9089-ee1e7a5a115e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.943761 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.948019 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0be92ca9-2a5c-4c33-b7e4-a440f157cd02","Type":"ContainerStarted","Data":"c53084eee5b7fe0bb8b05b829e8809ea409e48385e4fbd2a0ac9d27f46957f64"} Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.966023 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.9659956899999997 podStartE2EDuration="3.96599569s" podCreationTimestamp="2026-02-19 09:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:06.964121141 +0000 UTC m=+1168.591211409" watchObservedRunningTime="2026-02-19 09:03:06.96599569 +0000 UTC m=+1168.593085958" Feb 19 09:03:06 crc kubenswrapper[4675]: I0219 09:03:06.993586 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.001305 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.020792 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:07 crc kubenswrapper[4675]: E0219 09:03:07.022294 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-central-agent" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.022499 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-central-agent" Feb 19 09:03:07 crc kubenswrapper[4675]: E0219 09:03:07.022584 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="proxy-httpd" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.022745 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="proxy-httpd" Feb 19 09:03:07 crc kubenswrapper[4675]: E0219 09:03:07.023324 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-notification-agent" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.023382 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-notification-agent" Feb 19 09:03:07 crc kubenswrapper[4675]: E0219 09:03:07.023528 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="sg-core" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.023582 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="sg-core" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.023843 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-notification-agent" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.023935 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="sg-core" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.024008 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="proxy-httpd" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.024068 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" containerName="ceilometer-central-agent" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.025982 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.029855 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.032668 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.095554 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.135127 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5633bc1-203a-4dd9-9089-ee1e7a5a115e" path="/var/lib/kubelet/pods/d5633bc1-203a-4dd9-9089-ee1e7a5a115e/volumes" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.140475 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:07 crc kubenswrapper[4675]: E0219 09:03:07.141825 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-lmzr4 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.184910 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.184974 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-run-httpd\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.185002 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmzr4\" (UniqueName: \"kubernetes.io/projected/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-kube-api-access-lmzr4\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.185054 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-log-httpd\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.185126 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-scripts\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.185148 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-config-data\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.185185 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.286673 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-scripts\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.286734 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-config-data\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.286789 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.286818 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.286843 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-run-httpd\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.286867 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmzr4\" (UniqueName: \"kubernetes.io/projected/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-kube-api-access-lmzr4\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.286928 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-log-httpd\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.288598 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-log-httpd\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.290086 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-run-httpd\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.293980 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.294100 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-scripts\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.294900 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.300420 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-config-data\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.338997 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmzr4\" (UniqueName: \"kubernetes.io/projected/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-kube-api-access-lmzr4\") pod \"ceilometer-0\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.454007 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.454342 4675 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.468313 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.469207 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.598371 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-tls-certs\") pod \"d15338c6-aba7-4bec-a9c6-293e10903a7b\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.598470 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15338c6-aba7-4bec-a9c6-293e10903a7b-logs\") pod \"d15338c6-aba7-4bec-a9c6-293e10903a7b\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.598566 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-combined-ca-bundle\") pod \"d15338c6-aba7-4bec-a9c6-293e10903a7b\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.598611 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb675\" (UniqueName: \"kubernetes.io/projected/d15338c6-aba7-4bec-a9c6-293e10903a7b-kube-api-access-qb675\") pod \"d15338c6-aba7-4bec-a9c6-293e10903a7b\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.598740 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-scripts\") pod \"d15338c6-aba7-4bec-a9c6-293e10903a7b\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.598760 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-config-data\") pod \"d15338c6-aba7-4bec-a9c6-293e10903a7b\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.598791 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-secret-key\") pod \"d15338c6-aba7-4bec-a9c6-293e10903a7b\" (UID: \"d15338c6-aba7-4bec-a9c6-293e10903a7b\") " Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.599113 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d15338c6-aba7-4bec-a9c6-293e10903a7b-logs" (OuterVolumeSpecName: "logs") pod "d15338c6-aba7-4bec-a9c6-293e10903a7b" (UID: "d15338c6-aba7-4bec-a9c6-293e10903a7b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.599542 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15338c6-aba7-4bec-a9c6-293e10903a7b-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.609186 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15338c6-aba7-4bec-a9c6-293e10903a7b-kube-api-access-qb675" (OuterVolumeSpecName: "kube-api-access-qb675") pod "d15338c6-aba7-4bec-a9c6-293e10903a7b" (UID: "d15338c6-aba7-4bec-a9c6-293e10903a7b"). InnerVolumeSpecName "kube-api-access-qb675". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.611281 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d15338c6-aba7-4bec-a9c6-293e10903a7b" (UID: "d15338c6-aba7-4bec-a9c6-293e10903a7b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.633995 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d15338c6-aba7-4bec-a9c6-293e10903a7b" (UID: "d15338c6-aba7-4bec-a9c6-293e10903a7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.656515 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-scripts" (OuterVolumeSpecName: "scripts") pod "d15338c6-aba7-4bec-a9c6-293e10903a7b" (UID: "d15338c6-aba7-4bec-a9c6-293e10903a7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.656562 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-config-data" (OuterVolumeSpecName: "config-data") pod "d15338c6-aba7-4bec-a9c6-293e10903a7b" (UID: "d15338c6-aba7-4bec-a9c6-293e10903a7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.693902 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "d15338c6-aba7-4bec-a9c6-293e10903a7b" (UID: "d15338c6-aba7-4bec-a9c6-293e10903a7b"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.702981 4675 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.703014 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.703026 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb675\" (UniqueName: \"kubernetes.io/projected/d15338c6-aba7-4bec-a9c6-293e10903a7b-kube-api-access-qb675\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.703038 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.703046 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15338c6-aba7-4bec-a9c6-293e10903a7b-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.703056 4675 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15338c6-aba7-4bec-a9c6-293e10903a7b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.961187 4675 generic.go:334] "Generic (PLEG): container finished" podID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerID="56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff" exitCode=137 Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.961293 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bbbd6c858-v6gv9" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.961331 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bbbd6c858-v6gv9" event={"ID":"d15338c6-aba7-4bec-a9c6-293e10903a7b","Type":"ContainerDied","Data":"56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff"} Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.961380 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bbbd6c858-v6gv9" event={"ID":"d15338c6-aba7-4bec-a9c6-293e10903a7b","Type":"ContainerDied","Data":"b9b782a91e26d9785b05147cce47bc097bd4db31cf77a3e5cba4c3907d82264a"} Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.961400 4675 scope.go:117] "RemoveContainer" containerID="0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.961531 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:07 crc kubenswrapper[4675]: I0219 09:03:07.973759 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.004453 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bbbd6c858-v6gv9"] Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.018913 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bbbd6c858-v6gv9"] Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.109714 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-config-data\") pod \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.109783 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-combined-ca-bundle\") pod \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.110678 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-run-httpd\") pod \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.110811 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-scripts\") pod \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.110835 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-sg-core-conf-yaml\") pod \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.110963 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" (UID: "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.113225 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmzr4\" (UniqueName: \"kubernetes.io/projected/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-kube-api-access-lmzr4\") pod \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.113333 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-log-httpd\") pod \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\" (UID: \"4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e\") " Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.113792 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-config-data" (OuterVolumeSpecName: "config-data") pod "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" (UID: "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.114210 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" (UID: "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.114612 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-scripts" (OuterVolumeSpecName: "scripts") pod "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" (UID: "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.115753 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.115778 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.115788 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.115797 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.118219 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" (UID: "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.123958 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-kube-api-access-lmzr4" (OuterVolumeSpecName: "kube-api-access-lmzr4") pod "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" (UID: "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e"). InnerVolumeSpecName "kube-api-access-lmzr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.129754 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" (UID: "4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.158115 4675 scope.go:117] "RemoveContainer" containerID="56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.217773 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.217806 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmzr4\" (UniqueName: \"kubernetes.io/projected/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-kube-api-access-lmzr4\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.217817 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:08 crc kubenswrapper[4675]: I0219 09:03:08.973292 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.064890 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.075414 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.085127 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:09 crc kubenswrapper[4675]: E0219 09:03:09.085848 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon-log" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.085875 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon-log" Feb 19 09:03:09 crc kubenswrapper[4675]: E0219 09:03:09.085897 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.085908 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.086194 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon-log" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.086232 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" containerName="horizon" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.088912 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.093155 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.093626 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.102032 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.131673 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e" path="/var/lib/kubelet/pods/4c7f1b83-4295-4c84-a2a0-d4b4e58f4e3e/volumes" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.133380 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15338c6-aba7-4bec-a9c6-293e10903a7b" path="/var/lib/kubelet/pods/d15338c6-aba7-4bec-a9c6-293e10903a7b/volumes" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.134963 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-scripts\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.135113 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-config-data\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.135264 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-log-httpd\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.135399 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.135590 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.135711 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gjmg\" (UniqueName: \"kubernetes.io/projected/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-kube-api-access-7gjmg\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.135804 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-run-httpd\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.237208 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-log-httpd\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.237283 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.237441 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.237475 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gjmg\" (UniqueName: \"kubernetes.io/projected/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-kube-api-access-7gjmg\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.237503 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-run-httpd\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.237810 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-log-httpd\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.238092 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-run-httpd\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.238555 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-scripts\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.238713 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-config-data\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.245271 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-scripts\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.245573 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-config-data\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.248006 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.256923 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gjmg\" (UniqueName: \"kubernetes.io/projected/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-kube-api-access-7gjmg\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.263773 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " pod="openstack/ceilometer-0" Feb 19 09:03:09 crc kubenswrapper[4675]: I0219 09:03:09.424019 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:10 crc kubenswrapper[4675]: I0219 09:03:10.453113 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 19 09:03:13 crc kubenswrapper[4675]: I0219 09:03:13.240999 4675 scope.go:117] "RemoveContainer" containerID="0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2" Feb 19 09:03:13 crc kubenswrapper[4675]: E0219 09:03:13.242171 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2\": container with ID starting with 0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2 not found: ID does not exist" containerID="0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2" Feb 19 09:03:13 crc kubenswrapper[4675]: I0219 09:03:13.242204 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2"} err="failed to get container status \"0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2\": rpc error: code = NotFound desc = could not find container \"0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2\": container with ID starting with 0335ebfd9ae9b8f1beda3b173a6f8f6a9c6b1b742cc6a47d248b07be4e57c1c2 not found: ID does not exist" Feb 19 09:03:13 crc kubenswrapper[4675]: I0219 09:03:13.242245 4675 scope.go:117] "RemoveContainer" containerID="56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff" Feb 19 09:03:13 crc kubenswrapper[4675]: E0219 09:03:13.242722 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff\": container with ID starting with 56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff not found: ID does not exist" containerID="56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff" Feb 19 09:03:13 crc kubenswrapper[4675]: I0219 09:03:13.242749 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff"} err="failed to get container status \"56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff\": rpc error: code = NotFound desc = could not find container \"56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff\": container with ID starting with 56dc3dcce34328f63caa0ad9ced0a4bec11b68793d2fcfeb1c934181752d10ff not found: ID does not exist" Feb 19 09:03:13 crc kubenswrapper[4675]: I0219 09:03:13.376388 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:13 crc kubenswrapper[4675]: I0219 09:03:13.730420 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:13 crc kubenswrapper[4675]: W0219 09:03:13.735199 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ac9fa3f_b9f6_418b_8ac0_62e2ccdbc7bf.slice/crio-ca28b5c2d466f7f73e9d132315ffc9463aa085fefd38cba56f407f9b6ff2426e WatchSource:0}: Error finding container ca28b5c2d466f7f73e9d132315ffc9463aa085fefd38cba56f407f9b6ff2426e: Status 404 returned error can't find the container with id ca28b5c2d466f7f73e9d132315ffc9463aa085fefd38cba56f407f9b6ff2426e Feb 19 09:03:14 crc kubenswrapper[4675]: I0219 09:03:14.023028 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" event={"ID":"f6c638bb-53d1-4c7a-889b-a506b89b202e","Type":"ContainerStarted","Data":"0a3940db54abe60308aff0e453b5e139c01a362c8b43e4bc11a12997bbfa4822"} Feb 19 09:03:14 crc kubenswrapper[4675]: I0219 09:03:14.026836 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerStarted","Data":"ca28b5c2d466f7f73e9d132315ffc9463aa085fefd38cba56f407f9b6ff2426e"} Feb 19 09:03:14 crc kubenswrapper[4675]: I0219 09:03:14.042667 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" podStartSLOduration=2.410728127 podStartE2EDuration="12.04261784s" podCreationTimestamp="2026-02-19 09:03:02 +0000 UTC" firstStartedPulling="2026-02-19 09:03:03.677472891 +0000 UTC m=+1165.304563159" lastFinishedPulling="2026-02-19 09:03:13.309362604 +0000 UTC m=+1174.936452872" observedRunningTime="2026-02-19 09:03:14.038794581 +0000 UTC m=+1175.665884849" watchObservedRunningTime="2026-02-19 09:03:14.04261784 +0000 UTC m=+1175.669708108" Feb 19 09:03:14 crc kubenswrapper[4675]: I0219 09:03:14.290151 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 19 09:03:14 crc kubenswrapper[4675]: I0219 09:03:14.290218 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 19 09:03:14 crc kubenswrapper[4675]: I0219 09:03:14.331795 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 19 09:03:14 crc kubenswrapper[4675]: I0219 09:03:14.344851 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 19 09:03:15 crc kubenswrapper[4675]: I0219 09:03:15.036777 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerStarted","Data":"4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7"} Feb 19 09:03:15 crc kubenswrapper[4675]: I0219 09:03:15.037221 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 19 09:03:15 crc kubenswrapper[4675]: I0219 09:03:15.037461 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 19 09:03:16 crc kubenswrapper[4675]: I0219 09:03:16.047710 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerStarted","Data":"0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995"} Feb 19 09:03:16 crc kubenswrapper[4675]: I0219 09:03:16.048431 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerStarted","Data":"8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6"} Feb 19 09:03:17 crc kubenswrapper[4675]: I0219 09:03:17.018382 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 19 09:03:17 crc kubenswrapper[4675]: I0219 09:03:17.019844 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 19 09:03:18 crc kubenswrapper[4675]: I0219 09:03:18.070321 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerStarted","Data":"cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3"} Feb 19 09:03:18 crc kubenswrapper[4675]: I0219 09:03:18.070629 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-central-agent" containerID="cri-o://4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7" gracePeriod=30 Feb 19 09:03:18 crc kubenswrapper[4675]: I0219 09:03:18.070674 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="proxy-httpd" containerID="cri-o://cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3" gracePeriod=30 Feb 19 09:03:18 crc kubenswrapper[4675]: I0219 09:03:18.070701 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-notification-agent" containerID="cri-o://0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995" gracePeriod=30 Feb 19 09:03:18 crc kubenswrapper[4675]: I0219 09:03:18.070710 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="sg-core" containerID="cri-o://8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6" gracePeriod=30 Feb 19 09:03:18 crc kubenswrapper[4675]: I0219 09:03:18.099075 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.457341417 podStartE2EDuration="9.099045786s" podCreationTimestamp="2026-02-19 09:03:09 +0000 UTC" firstStartedPulling="2026-02-19 09:03:13.739297217 +0000 UTC m=+1175.366387485" lastFinishedPulling="2026-02-19 09:03:17.381001586 +0000 UTC m=+1179.008091854" observedRunningTime="2026-02-19 09:03:18.091820048 +0000 UTC m=+1179.718910316" watchObservedRunningTime="2026-02-19 09:03:18.099045786 +0000 UTC m=+1179.726136054" Feb 19 09:03:19 crc kubenswrapper[4675]: I0219 09:03:19.092905 4675 generic.go:334] "Generic (PLEG): container finished" podID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerID="cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3" exitCode=0 Feb 19 09:03:19 crc kubenswrapper[4675]: I0219 09:03:19.093276 4675 generic.go:334] "Generic (PLEG): container finished" podID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerID="8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6" exitCode=2 Feb 19 09:03:19 crc kubenswrapper[4675]: I0219 09:03:19.093287 4675 generic.go:334] "Generic (PLEG): container finished" podID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerID="0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995" exitCode=0 Feb 19 09:03:19 crc kubenswrapper[4675]: I0219 09:03:19.093003 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerDied","Data":"cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3"} Feb 19 09:03:19 crc kubenswrapper[4675]: I0219 09:03:19.093327 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerDied","Data":"8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6"} Feb 19 09:03:19 crc kubenswrapper[4675]: I0219 09:03:19.093345 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerDied","Data":"0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995"} Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.149579 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.150964 4675 generic.go:334] "Generic (PLEG): container finished" podID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerID="4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7" exitCode=0 Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.151361 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerDied","Data":"4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7"} Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.151397 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf","Type":"ContainerDied","Data":"ca28b5c2d466f7f73e9d132315ffc9463aa085fefd38cba56f407f9b6ff2426e"} Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.151418 4675 scope.go:117] "RemoveContainer" containerID="cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.172364 4675 scope.go:117] "RemoveContainer" containerID="8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.199353 4675 scope.go:117] "RemoveContainer" containerID="0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.228910 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-config-data\") pod \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.228974 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-sg-core-conf-yaml\") pod \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.229020 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-run-httpd\") pod \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.229058 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-log-httpd\") pod \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.229166 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-scripts\") pod \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.229220 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-combined-ca-bundle\") pod \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.229266 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gjmg\" (UniqueName: \"kubernetes.io/projected/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-kube-api-access-7gjmg\") pod \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\" (UID: \"0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf\") " Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.232455 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" (UID: "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.232620 4675 scope.go:117] "RemoveContainer" containerID="4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.233385 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" (UID: "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.237730 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-scripts" (OuterVolumeSpecName: "scripts") pod "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" (UID: "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.242497 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-kube-api-access-7gjmg" (OuterVolumeSpecName: "kube-api-access-7gjmg") pod "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" (UID: "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf"). InnerVolumeSpecName "kube-api-access-7gjmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.264020 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" (UID: "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.314050 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" (UID: "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.331428 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.331463 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.331475 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gjmg\" (UniqueName: \"kubernetes.io/projected/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-kube-api-access-7gjmg\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.331484 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.331494 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.331504 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.347819 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-config-data" (OuterVolumeSpecName: "config-data") pod "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" (UID: "0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.376301 4675 scope.go:117] "RemoveContainer" containerID="cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3" Feb 19 09:03:24 crc kubenswrapper[4675]: E0219 09:03:24.376848 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3\": container with ID starting with cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3 not found: ID does not exist" containerID="cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.376925 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3"} err="failed to get container status \"cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3\": rpc error: code = NotFound desc = could not find container \"cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3\": container with ID starting with cb247ec107b4a7acea750ac79e41a4cd4f7f01fdfe363d044fa8dcc335d2ccd3 not found: ID does not exist" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.376974 4675 scope.go:117] "RemoveContainer" containerID="8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6" Feb 19 09:03:24 crc kubenswrapper[4675]: E0219 09:03:24.377334 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6\": container with ID starting with 8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6 not found: ID does not exist" containerID="8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.377367 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6"} err="failed to get container status \"8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6\": rpc error: code = NotFound desc = could not find container \"8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6\": container with ID starting with 8314d126c10d44dce43913cfced88a3e5e625087d8ca681d055d9703b5f08ba6 not found: ID does not exist" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.377390 4675 scope.go:117] "RemoveContainer" containerID="0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995" Feb 19 09:03:24 crc kubenswrapper[4675]: E0219 09:03:24.377840 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995\": container with ID starting with 0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995 not found: ID does not exist" containerID="0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.377872 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995"} err="failed to get container status \"0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995\": rpc error: code = NotFound desc = could not find container \"0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995\": container with ID starting with 0791b2d2c67f11703f4d2f3cff0b145caff21c07c3e37e118032bd35f7224995 not found: ID does not exist" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.377893 4675 scope.go:117] "RemoveContainer" containerID="4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7" Feb 19 09:03:24 crc kubenswrapper[4675]: E0219 09:03:24.378178 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7\": container with ID starting with 4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7 not found: ID does not exist" containerID="4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.378208 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7"} err="failed to get container status \"4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7\": rpc error: code = NotFound desc = could not find container \"4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7\": container with ID starting with 4bf7368e8d720afeada1a249c95f5e0d8f87516d33b9ae521c861d326deffed7 not found: ID does not exist" Feb 19 09:03:24 crc kubenswrapper[4675]: I0219 09:03:24.436095 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.164171 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.190719 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.206586 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.217712 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:25 crc kubenswrapper[4675]: E0219 09:03:25.219066 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-central-agent" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219095 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-central-agent" Feb 19 09:03:25 crc kubenswrapper[4675]: E0219 09:03:25.219111 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="sg-core" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219120 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="sg-core" Feb 19 09:03:25 crc kubenswrapper[4675]: E0219 09:03:25.219133 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="proxy-httpd" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219139 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="proxy-httpd" Feb 19 09:03:25 crc kubenswrapper[4675]: E0219 09:03:25.219186 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-notification-agent" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219192 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-notification-agent" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219350 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-notification-agent" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219369 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="sg-core" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219377 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="ceilometer-central-agent" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.219388 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" containerName="proxy-httpd" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.221003 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.223054 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.225040 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.235939 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.248242 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8l94\" (UniqueName: \"kubernetes.io/projected/bac0550a-6f4d-478b-8f1b-4d86218a59e7-kube-api-access-b8l94\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.248315 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-scripts\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.248348 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-log-httpd\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.248420 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-run-httpd\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.248474 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.248506 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-config-data\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.248528 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.350076 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-scripts\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.350390 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-log-httpd\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.350503 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-run-httpd\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.350591 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.351467 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-config-data\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.351590 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.351814 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8l94\" (UniqueName: \"kubernetes.io/projected/bac0550a-6f4d-478b-8f1b-4d86218a59e7-kube-api-access-b8l94\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.351913 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-log-httpd\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.352294 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-run-httpd\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.357662 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.358429 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-scripts\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.358879 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-config-data\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.360425 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.370268 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8l94\" (UniqueName: \"kubernetes.io/projected/bac0550a-6f4d-478b-8f1b-4d86218a59e7-kube-api-access-b8l94\") pod \"ceilometer-0\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " pod="openstack/ceilometer-0" Feb 19 09:03:25 crc kubenswrapper[4675]: I0219 09:03:25.569185 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:26 crc kubenswrapper[4675]: I0219 09:03:26.043860 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:26 crc kubenswrapper[4675]: W0219 09:03:26.050045 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbac0550a_6f4d_478b_8f1b_4d86218a59e7.slice/crio-274cd1b5b01227e7f599a23f7828b1ffde30178e9f47e7dde9f76782081ac3ec WatchSource:0}: Error finding container 274cd1b5b01227e7f599a23f7828b1ffde30178e9f47e7dde9f76782081ac3ec: Status 404 returned error can't find the container with id 274cd1b5b01227e7f599a23f7828b1ffde30178e9f47e7dde9f76782081ac3ec Feb 19 09:03:26 crc kubenswrapper[4675]: I0219 09:03:26.175359 4675 generic.go:334] "Generic (PLEG): container finished" podID="f6c638bb-53d1-4c7a-889b-a506b89b202e" containerID="0a3940db54abe60308aff0e453b5e139c01a362c8b43e4bc11a12997bbfa4822" exitCode=0 Feb 19 09:03:26 crc kubenswrapper[4675]: I0219 09:03:26.176457 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" event={"ID":"f6c638bb-53d1-4c7a-889b-a506b89b202e","Type":"ContainerDied","Data":"0a3940db54abe60308aff0e453b5e139c01a362c8b43e4bc11a12997bbfa4822"} Feb 19 09:03:26 crc kubenswrapper[4675]: I0219 09:03:26.178265 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerStarted","Data":"274cd1b5b01227e7f599a23f7828b1ffde30178e9f47e7dde9f76782081ac3ec"} Feb 19 09:03:26 crc kubenswrapper[4675]: I0219 09:03:26.335768 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.116430 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf" path="/var/lib/kubelet/pods/0ac9fa3f-b9f6-418b-8ac0-62e2ccdbc7bf/volumes" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.193569 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerStarted","Data":"324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d"} Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.602382 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.638691 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j9vv\" (UniqueName: \"kubernetes.io/projected/f6c638bb-53d1-4c7a-889b-a506b89b202e-kube-api-access-2j9vv\") pod \"f6c638bb-53d1-4c7a-889b-a506b89b202e\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.638740 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-combined-ca-bundle\") pod \"f6c638bb-53d1-4c7a-889b-a506b89b202e\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.638771 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-scripts\") pod \"f6c638bb-53d1-4c7a-889b-a506b89b202e\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.638799 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-config-data\") pod \"f6c638bb-53d1-4c7a-889b-a506b89b202e\" (UID: \"f6c638bb-53d1-4c7a-889b-a506b89b202e\") " Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.645366 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-scripts" (OuterVolumeSpecName: "scripts") pod "f6c638bb-53d1-4c7a-889b-a506b89b202e" (UID: "f6c638bb-53d1-4c7a-889b-a506b89b202e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.645796 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6c638bb-53d1-4c7a-889b-a506b89b202e-kube-api-access-2j9vv" (OuterVolumeSpecName: "kube-api-access-2j9vv") pod "f6c638bb-53d1-4c7a-889b-a506b89b202e" (UID: "f6c638bb-53d1-4c7a-889b-a506b89b202e"). InnerVolumeSpecName "kube-api-access-2j9vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.669372 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6c638bb-53d1-4c7a-889b-a506b89b202e" (UID: "f6c638bb-53d1-4c7a-889b-a506b89b202e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.675696 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-config-data" (OuterVolumeSpecName: "config-data") pod "f6c638bb-53d1-4c7a-889b-a506b89b202e" (UID: "f6c638bb-53d1-4c7a-889b-a506b89b202e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.739915 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j9vv\" (UniqueName: \"kubernetes.io/projected/f6c638bb-53d1-4c7a-889b-a506b89b202e-kube-api-access-2j9vv\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.739955 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.739986 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:27 crc kubenswrapper[4675]: I0219 09:03:27.740003 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c638bb-53d1-4c7a-889b-a506b89b202e-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.205544 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" event={"ID":"f6c638bb-53d1-4c7a-889b-a506b89b202e","Type":"ContainerDied","Data":"905d2469db76fbd10e79379f73d1de371b3065bbe6677002fb5d3a67a559f247"} Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.205586 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="905d2469db76fbd10e79379f73d1de371b3065bbe6677002fb5d3a67a559f247" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.205618 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dlxqd" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.207602 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerStarted","Data":"2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56"} Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.207666 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerStarted","Data":"9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38"} Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.333633 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 19 09:03:28 crc kubenswrapper[4675]: E0219 09:03:28.334221 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6c638bb-53d1-4c7a-889b-a506b89b202e" containerName="nova-cell0-conductor-db-sync" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.334243 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6c638bb-53d1-4c7a-889b-a506b89b202e" containerName="nova-cell0-conductor-db-sync" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.334523 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6c638bb-53d1-4c7a-889b-a506b89b202e" containerName="nova-cell0-conductor-db-sync" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.335571 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.340032 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fgmtk" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.340415 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.349920 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408c0ffb-6a67-4351-ac43-8c5affe01d69-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.350084 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz6wc\" (UniqueName: \"kubernetes.io/projected/408c0ffb-6a67-4351-ac43-8c5affe01d69-kube-api-access-qz6wc\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.350125 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408c0ffb-6a67-4351-ac43-8c5affe01d69-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.350619 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.451016 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6wc\" (UniqueName: \"kubernetes.io/projected/408c0ffb-6a67-4351-ac43-8c5affe01d69-kube-api-access-qz6wc\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.451078 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408c0ffb-6a67-4351-ac43-8c5affe01d69-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.451183 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408c0ffb-6a67-4351-ac43-8c5affe01d69-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.456494 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408c0ffb-6a67-4351-ac43-8c5affe01d69-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.456984 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408c0ffb-6a67-4351-ac43-8c5affe01d69-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.469373 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz6wc\" (UniqueName: \"kubernetes.io/projected/408c0ffb-6a67-4351-ac43-8c5affe01d69-kube-api-access-qz6wc\") pod \"nova-cell0-conductor-0\" (UID: \"408c0ffb-6a67-4351-ac43-8c5affe01d69\") " pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:28 crc kubenswrapper[4675]: I0219 09:03:28.666863 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:29 crc kubenswrapper[4675]: I0219 09:03:29.126844 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 19 09:03:29 crc kubenswrapper[4675]: I0219 09:03:29.217817 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"408c0ffb-6a67-4351-ac43-8c5affe01d69","Type":"ContainerStarted","Data":"a67d198a44a4850bc632ba6e00f983fae805033adcd3bd50ee2b347486154f8c"} Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.230732 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerStarted","Data":"0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5"} Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.231116 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.230988 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="sg-core" containerID="cri-o://2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56" gracePeriod=30 Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.230916 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-central-agent" containerID="cri-o://324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d" gracePeriod=30 Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.231000 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-notification-agent" containerID="cri-o://9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38" gracePeriod=30 Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.231018 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="proxy-httpd" containerID="cri-o://0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5" gracePeriod=30 Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.234032 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"408c0ffb-6a67-4351-ac43-8c5affe01d69","Type":"ContainerStarted","Data":"ba6f9871a2d7a21c5f35a447ba1021cd3fe7de96110061a1bf5587a45f750bad"} Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.234908 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.264636 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.72855173 podStartE2EDuration="5.264611106s" podCreationTimestamp="2026-02-19 09:03:25 +0000 UTC" firstStartedPulling="2026-02-19 09:03:26.053979142 +0000 UTC m=+1187.681069410" lastFinishedPulling="2026-02-19 09:03:29.590038518 +0000 UTC m=+1191.217128786" observedRunningTime="2026-02-19 09:03:30.255249522 +0000 UTC m=+1191.882339790" watchObservedRunningTime="2026-02-19 09:03:30.264611106 +0000 UTC m=+1191.891701374" Feb 19 09:03:30 crc kubenswrapper[4675]: I0219 09:03:30.283963 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.28394275 podStartE2EDuration="2.28394275s" podCreationTimestamp="2026-02-19 09:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:30.273066126 +0000 UTC m=+1191.900156394" watchObservedRunningTime="2026-02-19 09:03:30.28394275 +0000 UTC m=+1191.911033018" Feb 19 09:03:31 crc kubenswrapper[4675]: I0219 09:03:31.246335 4675 generic.go:334] "Generic (PLEG): container finished" podID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerID="0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5" exitCode=0 Feb 19 09:03:31 crc kubenswrapper[4675]: I0219 09:03:31.246380 4675 generic.go:334] "Generic (PLEG): container finished" podID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerID="2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56" exitCode=2 Feb 19 09:03:31 crc kubenswrapper[4675]: I0219 09:03:31.246389 4675 generic.go:334] "Generic (PLEG): container finished" podID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerID="9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38" exitCode=0 Feb 19 09:03:31 crc kubenswrapper[4675]: I0219 09:03:31.246405 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerDied","Data":"0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5"} Feb 19 09:03:31 crc kubenswrapper[4675]: I0219 09:03:31.246456 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerDied","Data":"2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56"} Feb 19 09:03:31 crc kubenswrapper[4675]: I0219 09:03:31.246467 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerDied","Data":"9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38"} Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.217598 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.277330 4675 generic.go:334] "Generic (PLEG): container finished" podID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerID="324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d" exitCode=0 Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.277374 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerDied","Data":"324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d"} Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.277401 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bac0550a-6f4d-478b-8f1b-4d86218a59e7","Type":"ContainerDied","Data":"274cd1b5b01227e7f599a23f7828b1ffde30178e9f47e7dde9f76782081ac3ec"} Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.277418 4675 scope.go:117] "RemoveContainer" containerID="0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.277429 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.297897 4675 scope.go:117] "RemoveContainer" containerID="2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.316814 4675 scope.go:117] "RemoveContainer" containerID="9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.335947 4675 scope.go:117] "RemoveContainer" containerID="324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.353037 4675 scope.go:117] "RemoveContainer" containerID="0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5" Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.353471 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5\": container with ID starting with 0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5 not found: ID does not exist" containerID="0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.353499 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5"} err="failed to get container status \"0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5\": rpc error: code = NotFound desc = could not find container \"0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5\": container with ID starting with 0a4fd91e666bb614206f33146db809acc7043f400f81df15aa5b19f349a592e5 not found: ID does not exist" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.353520 4675 scope.go:117] "RemoveContainer" containerID="2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56" Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.353864 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56\": container with ID starting with 2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56 not found: ID does not exist" containerID="2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.353885 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56"} err="failed to get container status \"2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56\": rpc error: code = NotFound desc = could not find container \"2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56\": container with ID starting with 2b7e1c07c7569097eb61a33056248fd617a041a6f27390f7fe189e74431d0e56 not found: ID does not exist" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.353899 4675 scope.go:117] "RemoveContainer" containerID="9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38" Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.354274 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38\": container with ID starting with 9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38 not found: ID does not exist" containerID="9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.354297 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38"} err="failed to get container status \"9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38\": rpc error: code = NotFound desc = could not find container \"9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38\": container with ID starting with 9904bd81204c4a7d6c1fab4494c88863b17116057be62f761c9a3c81ff651e38 not found: ID does not exist" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.354311 4675 scope.go:117] "RemoveContainer" containerID="324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d" Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.354560 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d\": container with ID starting with 324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d not found: ID does not exist" containerID="324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.354580 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d"} err="failed to get container status \"324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d\": rpc error: code = NotFound desc = could not find container \"324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d\": container with ID starting with 324d1516582e2cbb45655d86706924e7d80272330c2f410691bd8ddb87d1e01d not found: ID does not exist" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.369361 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-sg-core-conf-yaml\") pod \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.369415 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-scripts\") pod \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.369521 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-log-httpd\") pod \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.369587 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-combined-ca-bundle\") pod \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.369654 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-run-httpd\") pod \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.369697 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-config-data\") pod \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.369721 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8l94\" (UniqueName: \"kubernetes.io/projected/bac0550a-6f4d-478b-8f1b-4d86218a59e7-kube-api-access-b8l94\") pod \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\" (UID: \"bac0550a-6f4d-478b-8f1b-4d86218a59e7\") " Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.370429 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bac0550a-6f4d-478b-8f1b-4d86218a59e7" (UID: "bac0550a-6f4d-478b-8f1b-4d86218a59e7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.370438 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bac0550a-6f4d-478b-8f1b-4d86218a59e7" (UID: "bac0550a-6f4d-478b-8f1b-4d86218a59e7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.375285 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-scripts" (OuterVolumeSpecName: "scripts") pod "bac0550a-6f4d-478b-8f1b-4d86218a59e7" (UID: "bac0550a-6f4d-478b-8f1b-4d86218a59e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.375723 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bac0550a-6f4d-478b-8f1b-4d86218a59e7-kube-api-access-b8l94" (OuterVolumeSpecName: "kube-api-access-b8l94") pod "bac0550a-6f4d-478b-8f1b-4d86218a59e7" (UID: "bac0550a-6f4d-478b-8f1b-4d86218a59e7"). InnerVolumeSpecName "kube-api-access-b8l94". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.398401 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bac0550a-6f4d-478b-8f1b-4d86218a59e7" (UID: "bac0550a-6f4d-478b-8f1b-4d86218a59e7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.453820 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bac0550a-6f4d-478b-8f1b-4d86218a59e7" (UID: "bac0550a-6f4d-478b-8f1b-4d86218a59e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.472199 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.472231 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.472242 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bac0550a-6f4d-478b-8f1b-4d86218a59e7-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.472254 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8l94\" (UniqueName: \"kubernetes.io/projected/bac0550a-6f4d-478b-8f1b-4d86218a59e7-kube-api-access-b8l94\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.472263 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.472271 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.472634 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-config-data" (OuterVolumeSpecName: "config-data") pod "bac0550a-6f4d-478b-8f1b-4d86218a59e7" (UID: "bac0550a-6f4d-478b-8f1b-4d86218a59e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.574448 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bac0550a-6f4d-478b-8f1b-4d86218a59e7-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.621347 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.631486 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.641508 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.641937 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-notification-agent" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.641958 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-notification-agent" Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.641974 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="proxy-httpd" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.641981 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="proxy-httpd" Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.642007 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="sg-core" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.642013 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="sg-core" Feb 19 09:03:34 crc kubenswrapper[4675]: E0219 09:03:34.642022 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-central-agent" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.642029 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-central-agent" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.642199 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="proxy-httpd" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.642224 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-notification-agent" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.642233 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="sg-core" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.642246 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" containerName="ceilometer-central-agent" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.644219 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.645899 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.646615 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.668391 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.780572 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-log-httpd\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.780686 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65gfj\" (UniqueName: \"kubernetes.io/projected/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-kube-api-access-65gfj\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.780875 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.780905 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.780947 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-run-httpd\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.780987 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-config-data\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.781098 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-scripts\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883085 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-log-httpd\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883156 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65gfj\" (UniqueName: \"kubernetes.io/projected/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-kube-api-access-65gfj\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883226 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883261 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883293 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-run-httpd\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883325 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-config-data\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883402 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-scripts\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883756 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-log-httpd\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.883824 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-run-httpd\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.889209 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.889402 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.889424 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-scripts\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.890302 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-config-data\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.899979 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65gfj\" (UniqueName: \"kubernetes.io/projected/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-kube-api-access-65gfj\") pod \"ceilometer-0\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " pod="openstack/ceilometer-0" Feb 19 09:03:34 crc kubenswrapper[4675]: I0219 09:03:34.961195 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:03:35 crc kubenswrapper[4675]: I0219 09:03:35.115481 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bac0550a-6f4d-478b-8f1b-4d86218a59e7" path="/var/lib/kubelet/pods/bac0550a-6f4d-478b-8f1b-4d86218a59e7/volumes" Feb 19 09:03:35 crc kubenswrapper[4675]: I0219 09:03:35.421857 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:03:36 crc kubenswrapper[4675]: I0219 09:03:36.299879 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerStarted","Data":"350a30b2c1263a9f0f532edc6181d31023c405138a6de80b79c7e4ef6cbb5f2e"} Feb 19 09:03:37 crc kubenswrapper[4675]: I0219 09:03:37.309329 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerStarted","Data":"8fd247fc8e94099dde6448180737030d5004728475f60b8fdde291f8835ec9f2"} Feb 19 09:03:37 crc kubenswrapper[4675]: I0219 09:03:37.311066 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerStarted","Data":"98caa5676c91647f6dcd5bd202334094f163141ac453c1f606241c1c3c04927f"} Feb 19 09:03:38 crc kubenswrapper[4675]: I0219 09:03:38.319582 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerStarted","Data":"2ce9eddfa923ad9b9a7e17f16a704c8123f9ba55ae63b2505fd9dc9368bf307d"} Feb 19 09:03:38 crc kubenswrapper[4675]: I0219 09:03:38.699784 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.173449 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fjxzh"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.175622 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.177992 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.184098 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.224228 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fjxzh"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.312476 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.312543 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-scripts\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.312735 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-config-data\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.312756 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v82f\" (UniqueName: \"kubernetes.io/projected/5077fc1f-d1da-4683-84e8-52afdefcd0e1-kube-api-access-8v82f\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.419331 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-config-data\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.419366 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v82f\" (UniqueName: \"kubernetes.io/projected/5077fc1f-d1da-4683-84e8-52afdefcd0e1-kube-api-access-8v82f\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.419403 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.419429 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-scripts\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.424558 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.425840 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.427651 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.430183 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-scripts\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.432367 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-config-data\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.437502 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.444441 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v82f\" (UniqueName: \"kubernetes.io/projected/5077fc1f-d1da-4683-84e8-52afdefcd0e1-kube-api-access-8v82f\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.475752 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fjxzh\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.504222 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.505968 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.510777 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.553624 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.555499 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.557313 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.576785 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.580979 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.590196 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.627139 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.627196 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.627239 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjvrc\" (UniqueName: \"kubernetes.io/projected/8969cd9a-d085-42c8-910d-8cb09bbeb93b-kube-api-access-wjvrc\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.627293 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-config-data\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.627321 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.627419 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8969cd9a-d085-42c8-910d-8cb09bbeb93b-logs\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.627492 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bswpj\" (UniqueName: \"kubernetes.io/projected/15194877-7f24-46c5-aab3-229f004a3b7d-kube-api-access-bswpj\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.642249 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.643886 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.651128 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.652003 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.691217 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vbjhx"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.694259 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.727633 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vbjhx"] Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.728873 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8969cd9a-d085-42c8-910d-8cb09bbeb93b-logs\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.728944 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bswpj\" (UniqueName: \"kubernetes.io/projected/15194877-7f24-46c5-aab3-229f004a3b7d-kube-api-access-bswpj\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.728980 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jk5f\" (UniqueName: \"kubernetes.io/projected/2e84d98e-c6f1-4551-9e45-7085e31720ae-kube-api-access-7jk5f\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729017 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729035 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729054 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-config-data\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729080 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjvrc\" (UniqueName: \"kubernetes.io/projected/8969cd9a-d085-42c8-910d-8cb09bbeb93b-kube-api-access-wjvrc\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729122 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-config-data\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729145 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729202 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.729408 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8969cd9a-d085-42c8-910d-8cb09bbeb93b-logs\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.737160 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-config-data\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.738334 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.748840 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.749358 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.756483 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjvrc\" (UniqueName: \"kubernetes.io/projected/8969cd9a-d085-42c8-910d-8cb09bbeb93b-kube-api-access-wjvrc\") pod \"nova-api-0\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.761368 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bswpj\" (UniqueName: \"kubernetes.io/projected/15194877-7f24-46c5-aab3-229f004a3b7d-kube-api-access-bswpj\") pod \"nova-cell1-novncproxy-0\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830768 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830826 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-config\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830852 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5zxc\" (UniqueName: \"kubernetes.io/projected/f5a1bc99-3e8c-4078-a36f-853c6f785137-kube-api-access-f5zxc\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830875 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830902 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563a58c4-8996-4ff3-993d-417e6023922f-logs\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830929 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-config-data\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830952 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.830973 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-svc\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.831004 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.831070 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twwnt\" (UniqueName: \"kubernetes.io/projected/563a58c4-8996-4ff3-993d-417e6023922f-kube-api-access-twwnt\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.831109 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jk5f\" (UniqueName: \"kubernetes.io/projected/2e84d98e-c6f1-4551-9e45-7085e31720ae-kube-api-access-7jk5f\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.831132 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.831155 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-config-data\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.842135 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-config-data\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.844011 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.859309 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jk5f\" (UniqueName: \"kubernetes.io/projected/2e84d98e-c6f1-4551-9e45-7085e31720ae-kube-api-access-7jk5f\") pod \"nova-scheduler-0\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.877245 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.899614 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.911500 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.942893 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5zxc\" (UniqueName: \"kubernetes.io/projected/f5a1bc99-3e8c-4078-a36f-853c6f785137-kube-api-access-f5zxc\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.942979 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.943026 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563a58c4-8996-4ff3-993d-417e6023922f-logs\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.943053 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-config-data\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.944098 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.944163 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-svc\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.944343 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.945509 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twwnt\" (UniqueName: \"kubernetes.io/projected/563a58c4-8996-4ff3-993d-417e6023922f-kube-api-access-twwnt\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.945621 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.945733 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.945780 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-config\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.946586 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-config\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.946939 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.947315 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563a58c4-8996-4ff3-993d-417e6023922f-logs\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.947782 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-svc\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.948127 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.959296 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-config-data\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.964442 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.969607 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twwnt\" (UniqueName: \"kubernetes.io/projected/563a58c4-8996-4ff3-993d-417e6023922f-kube-api-access-twwnt\") pod \"nova-metadata-0\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.989160 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:39 crc kubenswrapper[4675]: I0219 09:03:39.990410 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5zxc\" (UniqueName: \"kubernetes.io/projected/f5a1bc99-3e8c-4078-a36f-853c6f785137-kube-api-access-f5zxc\") pod \"dnsmasq-dns-865f5d856f-vbjhx\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.067039 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.262182 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fjxzh"] Feb 19 09:03:40 crc kubenswrapper[4675]: W0219 09:03:40.353830 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5077fc1f_d1da_4683_84e8_52afdefcd0e1.slice/crio-7047a8beb50cd9e179bbcd78227bdbfa80928099ac16dabdf2dfd46f0ac043a6 WatchSource:0}: Error finding container 7047a8beb50cd9e179bbcd78227bdbfa80928099ac16dabdf2dfd46f0ac043a6: Status 404 returned error can't find the container with id 7047a8beb50cd9e179bbcd78227bdbfa80928099ac16dabdf2dfd46f0ac043a6 Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.389278 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fjxzh" event={"ID":"5077fc1f-d1da-4683-84e8-52afdefcd0e1","Type":"ContainerStarted","Data":"7047a8beb50cd9e179bbcd78227bdbfa80928099ac16dabdf2dfd46f0ac043a6"} Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.407878 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerStarted","Data":"9dac50a6e901cab039ecef2c07f3d1c043af5dccbc418fcecefedccc70f3ca98"} Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.408075 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.450109 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8207197539999997 podStartE2EDuration="6.450079531s" podCreationTimestamp="2026-02-19 09:03:34 +0000 UTC" firstStartedPulling="2026-02-19 09:03:35.430743676 +0000 UTC m=+1197.057833944" lastFinishedPulling="2026-02-19 09:03:39.060103453 +0000 UTC m=+1200.687193721" observedRunningTime="2026-02-19 09:03:40.443396247 +0000 UTC m=+1202.070486515" watchObservedRunningTime="2026-02-19 09:03:40.450079531 +0000 UTC m=+1202.077169809" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.629129 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.747444 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8lmq6"] Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.748885 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.756432 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.756615 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.795932 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8lmq6"] Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.880742 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.880873 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-scripts\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.880943 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6smfg\" (UniqueName: \"kubernetes.io/projected/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-kube-api-access-6smfg\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.880972 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.934720 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.984816 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.984901 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-scripts\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.984960 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6smfg\" (UniqueName: \"kubernetes.io/projected/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-kube-api-access-6smfg\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.984984 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.990629 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-scripts\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.990785 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:40 crc kubenswrapper[4675]: I0219 09:03:40.992394 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.000819 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6smfg\" (UniqueName: \"kubernetes.io/projected/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-kube-api-access-6smfg\") pod \"nova-cell1-conductor-db-sync-8lmq6\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.096081 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.136184 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.172790 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.202112 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vbjhx"] Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.434475 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8969cd9a-d085-42c8-910d-8cb09bbeb93b","Type":"ContainerStarted","Data":"965133f30b80317846b491770f4a04cb5119176eb5f1e5e1e433ecfa4107c685"} Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.464777 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"563a58c4-8996-4ff3-993d-417e6023922f","Type":"ContainerStarted","Data":"604326a8c02ae1d66e8c5891ebcb6251c1d04b2b6134174bde38d14e997c9412"} Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.475687 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e84d98e-c6f1-4551-9e45-7085e31720ae","Type":"ContainerStarted","Data":"0baaa3919cef179f32913aee966ba3dc6ca43583884b1d3a9837b33e35916047"} Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.479679 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" event={"ID":"f5a1bc99-3e8c-4078-a36f-853c6f785137","Type":"ContainerStarted","Data":"bfb83dbdd46f546da01009b0f26ce051e727d6536965774b0ff8750ccbc89441"} Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.481044 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fjxzh" event={"ID":"5077fc1f-d1da-4683-84e8-52afdefcd0e1","Type":"ContainerStarted","Data":"8405c98fd4eab583741f22c9bcfcf596e2dd54cdbf4d0b8ecca8e43e6115109d"} Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.484768 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15194877-7f24-46c5-aab3-229f004a3b7d","Type":"ContainerStarted","Data":"bb9a05df624375bf0278d6c9b8382ee8d7ab4f5f2a68c571fc360d98264ad1f8"} Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.504745 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fjxzh" podStartSLOduration=2.504720651 podStartE2EDuration="2.504720651s" podCreationTimestamp="2026-02-19 09:03:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:41.496040445 +0000 UTC m=+1203.123130973" watchObservedRunningTime="2026-02-19 09:03:41.504720651 +0000 UTC m=+1203.131810949" Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.735385 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8lmq6"] Feb 19 09:03:41 crc kubenswrapper[4675]: W0219 09:03:41.741510 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f455bdc_3fe8_4dd4_88e8_a1a24a7f6102.slice/crio-9659fa92154b85665be4657d357ae5bf908ba64f59325b0f10275ef162156620 WatchSource:0}: Error finding container 9659fa92154b85665be4657d357ae5bf908ba64f59325b0f10275ef162156620: Status 404 returned error can't find the container with id 9659fa92154b85665be4657d357ae5bf908ba64f59325b0f10275ef162156620 Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.743868 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:03:41 crc kubenswrapper[4675]: I0219 09:03:41.743947 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:03:42 crc kubenswrapper[4675]: I0219 09:03:42.494933 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" event={"ID":"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102","Type":"ContainerStarted","Data":"3e3ce76b3c32778f16f95ff346231639cc54324dd25df20f79a0f043e8bd655a"} Feb 19 09:03:42 crc kubenswrapper[4675]: I0219 09:03:42.495319 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" event={"ID":"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102","Type":"ContainerStarted","Data":"9659fa92154b85665be4657d357ae5bf908ba64f59325b0f10275ef162156620"} Feb 19 09:03:42 crc kubenswrapper[4675]: I0219 09:03:42.498159 4675 generic.go:334] "Generic (PLEG): container finished" podID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerID="adbe0d9e78fcc385581c1b309f49a120d8f2e65e17554a1897145951dece975f" exitCode=0 Feb 19 09:03:42 crc kubenswrapper[4675]: I0219 09:03:42.498219 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" event={"ID":"f5a1bc99-3e8c-4078-a36f-853c6f785137","Type":"ContainerDied","Data":"adbe0d9e78fcc385581c1b309f49a120d8f2e65e17554a1897145951dece975f"} Feb 19 09:03:42 crc kubenswrapper[4675]: I0219 09:03:42.513893 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" podStartSLOduration=2.513867935 podStartE2EDuration="2.513867935s" podCreationTimestamp="2026-02-19 09:03:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:42.51022403 +0000 UTC m=+1204.137314298" watchObservedRunningTime="2026-02-19 09:03:42.513867935 +0000 UTC m=+1204.140958203" Feb 19 09:03:43 crc kubenswrapper[4675]: I0219 09:03:43.326006 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:03:43 crc kubenswrapper[4675]: I0219 09:03:43.413786 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.535940 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e84d98e-c6f1-4551-9e45-7085e31720ae","Type":"ContainerStarted","Data":"1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9"} Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.538268 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" event={"ID":"f5a1bc99-3e8c-4078-a36f-853c6f785137","Type":"ContainerStarted","Data":"1eb8570789ed39dac9f15de3c04ad5cdd9e3a27049c73a3af740861c59ac7368"} Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.538405 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.539743 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15194877-7f24-46c5-aab3-229f004a3b7d","Type":"ContainerStarted","Data":"8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9"} Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.539831 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="15194877-7f24-46c5-aab3-229f004a3b7d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9" gracePeriod=30 Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.543059 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8969cd9a-d085-42c8-910d-8cb09bbeb93b","Type":"ContainerStarted","Data":"300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79"} Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.543104 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8969cd9a-d085-42c8-910d-8cb09bbeb93b","Type":"ContainerStarted","Data":"202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c"} Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.545259 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"563a58c4-8996-4ff3-993d-417e6023922f","Type":"ContainerStarted","Data":"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807"} Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.545288 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"563a58c4-8996-4ff3-993d-417e6023922f","Type":"ContainerStarted","Data":"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884"} Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.545342 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-log" containerID="cri-o://4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884" gracePeriod=30 Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.545364 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-metadata" containerID="cri-o://38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807" gracePeriod=30 Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.586519 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.131067402 podStartE2EDuration="7.586501863s" podCreationTimestamp="2026-02-19 09:03:39 +0000 UTC" firstStartedPulling="2026-02-19 09:03:41.227062887 +0000 UTC m=+1202.854153165" lastFinishedPulling="2026-02-19 09:03:45.682497358 +0000 UTC m=+1207.309587626" observedRunningTime="2026-02-19 09:03:46.578212577 +0000 UTC m=+1208.205302845" watchObservedRunningTime="2026-02-19 09:03:46.586501863 +0000 UTC m=+1208.213592131" Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.586890 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.085978117 podStartE2EDuration="7.586885183s" podCreationTimestamp="2026-02-19 09:03:39 +0000 UTC" firstStartedPulling="2026-02-19 09:03:41.182094755 +0000 UTC m=+1202.809185023" lastFinishedPulling="2026-02-19 09:03:45.683001821 +0000 UTC m=+1207.310092089" observedRunningTime="2026-02-19 09:03:46.558158545 +0000 UTC m=+1208.185248813" watchObservedRunningTime="2026-02-19 09:03:46.586885183 +0000 UTC m=+1208.213975441" Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.606552 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" podStartSLOduration=7.606533635 podStartE2EDuration="7.606533635s" podCreationTimestamp="2026-02-19 09:03:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:46.598887356 +0000 UTC m=+1208.225977624" watchObservedRunningTime="2026-02-19 09:03:46.606533635 +0000 UTC m=+1208.233623903" Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.634369 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.598950787 podStartE2EDuration="7.6343464s" podCreationTimestamp="2026-02-19 09:03:39 +0000 UTC" firstStartedPulling="2026-02-19 09:03:40.651474179 +0000 UTC m=+1202.278564447" lastFinishedPulling="2026-02-19 09:03:45.686869792 +0000 UTC m=+1207.313960060" observedRunningTime="2026-02-19 09:03:46.631079525 +0000 UTC m=+1208.258169783" watchObservedRunningTime="2026-02-19 09:03:46.6343464 +0000 UTC m=+1208.261436678" Feb 19 09:03:46 crc kubenswrapper[4675]: I0219 09:03:46.641317 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.893372898 podStartE2EDuration="7.64128065s" podCreationTimestamp="2026-02-19 09:03:39 +0000 UTC" firstStartedPulling="2026-02-19 09:03:40.933295272 +0000 UTC m=+1202.560385540" lastFinishedPulling="2026-02-19 09:03:45.681203024 +0000 UTC m=+1207.308293292" observedRunningTime="2026-02-19 09:03:46.619585715 +0000 UTC m=+1208.246675993" watchObservedRunningTime="2026-02-19 09:03:46.64128065 +0000 UTC m=+1208.268370918" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.213517 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.353767 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-config-data\") pod \"563a58c4-8996-4ff3-993d-417e6023922f\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.354007 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twwnt\" (UniqueName: \"kubernetes.io/projected/563a58c4-8996-4ff3-993d-417e6023922f-kube-api-access-twwnt\") pod \"563a58c4-8996-4ff3-993d-417e6023922f\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.354143 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563a58c4-8996-4ff3-993d-417e6023922f-logs\") pod \"563a58c4-8996-4ff3-993d-417e6023922f\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.354180 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-combined-ca-bundle\") pod \"563a58c4-8996-4ff3-993d-417e6023922f\" (UID: \"563a58c4-8996-4ff3-993d-417e6023922f\") " Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.354971 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/563a58c4-8996-4ff3-993d-417e6023922f-logs" (OuterVolumeSpecName: "logs") pod "563a58c4-8996-4ff3-993d-417e6023922f" (UID: "563a58c4-8996-4ff3-993d-417e6023922f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.456495 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563a58c4-8996-4ff3-993d-417e6023922f-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.554929 4675 generic.go:334] "Generic (PLEG): container finished" podID="563a58c4-8996-4ff3-993d-417e6023922f" containerID="38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807" exitCode=0 Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.554970 4675 generic.go:334] "Generic (PLEG): container finished" podID="563a58c4-8996-4ff3-993d-417e6023922f" containerID="4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884" exitCode=143 Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.554983 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.555053 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"563a58c4-8996-4ff3-993d-417e6023922f","Type":"ContainerDied","Data":"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807"} Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.555123 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"563a58c4-8996-4ff3-993d-417e6023922f","Type":"ContainerDied","Data":"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884"} Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.555140 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"563a58c4-8996-4ff3-993d-417e6023922f","Type":"ContainerDied","Data":"604326a8c02ae1d66e8c5891ebcb6251c1d04b2b6134174bde38d14e997c9412"} Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.555161 4675 scope.go:117] "RemoveContainer" containerID="38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.595612 4675 scope.go:117] "RemoveContainer" containerID="4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.620285 4675 scope.go:117] "RemoveContainer" containerID="38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807" Feb 19 09:03:47 crc kubenswrapper[4675]: E0219 09:03:47.621102 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807\": container with ID starting with 38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807 not found: ID does not exist" containerID="38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.621155 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807"} err="failed to get container status \"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807\": rpc error: code = NotFound desc = could not find container \"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807\": container with ID starting with 38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807 not found: ID does not exist" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.621186 4675 scope.go:117] "RemoveContainer" containerID="4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884" Feb 19 09:03:47 crc kubenswrapper[4675]: E0219 09:03:47.621658 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884\": container with ID starting with 4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884 not found: ID does not exist" containerID="4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.621707 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884"} err="failed to get container status \"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884\": rpc error: code = NotFound desc = could not find container \"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884\": container with ID starting with 4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884 not found: ID does not exist" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.621735 4675 scope.go:117] "RemoveContainer" containerID="38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.622050 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807"} err="failed to get container status \"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807\": rpc error: code = NotFound desc = could not find container \"38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807\": container with ID starting with 38c7233751dc848fcbebdebe7d3284f7f0e80038622a4e61af5d0a7f1c8d0807 not found: ID does not exist" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.622084 4675 scope.go:117] "RemoveContainer" containerID="4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.622369 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884"} err="failed to get container status \"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884\": rpc error: code = NotFound desc = could not find container \"4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884\": container with ID starting with 4360408d27b1989e6c26fb8aae652c01f44cac95ade4fadabe09ebdbea9f5884 not found: ID does not exist" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.635897 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563a58c4-8996-4ff3-993d-417e6023922f-kube-api-access-twwnt" (OuterVolumeSpecName: "kube-api-access-twwnt") pod "563a58c4-8996-4ff3-993d-417e6023922f" (UID: "563a58c4-8996-4ff3-993d-417e6023922f"). InnerVolumeSpecName "kube-api-access-twwnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.643612 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-config-data" (OuterVolumeSpecName: "config-data") pod "563a58c4-8996-4ff3-993d-417e6023922f" (UID: "563a58c4-8996-4ff3-993d-417e6023922f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.647578 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "563a58c4-8996-4ff3-993d-417e6023922f" (UID: "563a58c4-8996-4ff3-993d-417e6023922f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.659008 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.659043 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563a58c4-8996-4ff3-993d-417e6023922f-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.659052 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twwnt\" (UniqueName: \"kubernetes.io/projected/563a58c4-8996-4ff3-993d-417e6023922f-kube-api-access-twwnt\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.896030 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.911410 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.930257 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:47 crc kubenswrapper[4675]: E0219 09:03:47.930900 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-log" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.930925 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-log" Feb 19 09:03:47 crc kubenswrapper[4675]: E0219 09:03:47.930940 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-metadata" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.930949 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-metadata" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.931178 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-metadata" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.931206 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="563a58c4-8996-4ff3-993d-417e6023922f" containerName="nova-metadata-log" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.932479 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.934692 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.934735 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 19 09:03:47 crc kubenswrapper[4675]: I0219 09:03:47.941491 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.068277 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vdmg\" (UniqueName: \"kubernetes.io/projected/8d772af7-a796-4c1a-af7d-04c440a68d86-kube-api-access-2vdmg\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.068352 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-config-data\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.068387 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.068459 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.068936 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d772af7-a796-4c1a-af7d-04c440a68d86-logs\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.171670 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vdmg\" (UniqueName: \"kubernetes.io/projected/8d772af7-a796-4c1a-af7d-04c440a68d86-kube-api-access-2vdmg\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.171749 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-config-data\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.171783 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.171812 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.171884 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d772af7-a796-4c1a-af7d-04c440a68d86-logs\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.172411 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d772af7-a796-4c1a-af7d-04c440a68d86-logs\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.178424 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.178839 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-config-data\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.188389 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.192202 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vdmg\" (UniqueName: \"kubernetes.io/projected/8d772af7-a796-4c1a-af7d-04c440a68d86-kube-api-access-2vdmg\") pod \"nova-metadata-0\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.261409 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:48 crc kubenswrapper[4675]: I0219 09:03:48.703885 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.133196 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="563a58c4-8996-4ff3-993d-417e6023922f" path="/var/lib/kubelet/pods/563a58c4-8996-4ff3-993d-417e6023922f/volumes" Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.576117 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d772af7-a796-4c1a-af7d-04c440a68d86","Type":"ContainerStarted","Data":"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac"} Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.576449 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d772af7-a796-4c1a-af7d-04c440a68d86","Type":"ContainerStarted","Data":"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c"} Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.576463 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d772af7-a796-4c1a-af7d-04c440a68d86","Type":"ContainerStarted","Data":"bd2010c8a72dbd377d1494af9d10620cea6c34379728ce5931416383a43a93e5"} Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.578323 4675 generic.go:334] "Generic (PLEG): container finished" podID="5077fc1f-d1da-4683-84e8-52afdefcd0e1" containerID="8405c98fd4eab583741f22c9bcfcf596e2dd54cdbf4d0b8ecca8e43e6115109d" exitCode=0 Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.578382 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fjxzh" event={"ID":"5077fc1f-d1da-4683-84e8-52afdefcd0e1","Type":"ContainerDied","Data":"8405c98fd4eab583741f22c9bcfcf596e2dd54cdbf4d0b8ecca8e43e6115109d"} Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.597799 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.597779596 podStartE2EDuration="2.597779596s" podCreationTimestamp="2026-02-19 09:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:49.593091503 +0000 UTC m=+1211.220181771" watchObservedRunningTime="2026-02-19 09:03:49.597779596 +0000 UTC m=+1211.224869864" Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.878164 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.899879 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.899931 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.912753 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.912823 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 19 09:03:49 crc kubenswrapper[4675]: I0219 09:03:49.942764 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 19 09:03:50 crc kubenswrapper[4675]: I0219 09:03:50.621010 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 19 09:03:50 crc kubenswrapper[4675]: I0219 09:03:50.941294 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:03:50 crc kubenswrapper[4675]: I0219 09:03:50.941335 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:03:50 crc kubenswrapper[4675]: I0219 09:03:50.979082 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.123787 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v82f\" (UniqueName: \"kubernetes.io/projected/5077fc1f-d1da-4683-84e8-52afdefcd0e1-kube-api-access-8v82f\") pod \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.123913 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-config-data\") pod \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.123954 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-scripts\") pod \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.124192 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-combined-ca-bundle\") pod \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\" (UID: \"5077fc1f-d1da-4683-84e8-52afdefcd0e1\") " Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.132410 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-scripts" (OuterVolumeSpecName: "scripts") pod "5077fc1f-d1da-4683-84e8-52afdefcd0e1" (UID: "5077fc1f-d1da-4683-84e8-52afdefcd0e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.142873 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5077fc1f-d1da-4683-84e8-52afdefcd0e1-kube-api-access-8v82f" (OuterVolumeSpecName: "kube-api-access-8v82f") pod "5077fc1f-d1da-4683-84e8-52afdefcd0e1" (UID: "5077fc1f-d1da-4683-84e8-52afdefcd0e1"). InnerVolumeSpecName "kube-api-access-8v82f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.161970 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-config-data" (OuterVolumeSpecName: "config-data") pod "5077fc1f-d1da-4683-84e8-52afdefcd0e1" (UID: "5077fc1f-d1da-4683-84e8-52afdefcd0e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.175797 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5077fc1f-d1da-4683-84e8-52afdefcd0e1" (UID: "5077fc1f-d1da-4683-84e8-52afdefcd0e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.227224 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.227610 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.227625 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5077fc1f-d1da-4683-84e8-52afdefcd0e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.227677 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v82f\" (UniqueName: \"kubernetes.io/projected/5077fc1f-d1da-4683-84e8-52afdefcd0e1-kube-api-access-8v82f\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.599939 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fjxzh" event={"ID":"5077fc1f-d1da-4683-84e8-52afdefcd0e1","Type":"ContainerDied","Data":"7047a8beb50cd9e179bbcd78227bdbfa80928099ac16dabdf2dfd46f0ac043a6"} Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.599992 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fjxzh" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.599992 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7047a8beb50cd9e179bbcd78227bdbfa80928099ac16dabdf2dfd46f0ac043a6" Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.792708 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.792933 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-log" containerID="cri-o://202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c" gracePeriod=30 Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.792998 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-api" containerID="cri-o://300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79" gracePeriod=30 Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.809922 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.938344 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.938814 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-log" containerID="cri-o://812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c" gracePeriod=30 Feb 19 09:03:51 crc kubenswrapper[4675]: I0219 09:03:51.938901 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-metadata" containerID="cri-o://5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac" gracePeriod=30 Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.611757 4675 generic.go:334] "Generic (PLEG): container finished" podID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerID="202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c" exitCode=143 Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.612304 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8969cd9a-d085-42c8-910d-8cb09bbeb93b","Type":"ContainerDied","Data":"202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c"} Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.613901 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.615010 4675 generic.go:334] "Generic (PLEG): container finished" podID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerID="5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac" exitCode=0 Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.615037 4675 generic.go:334] "Generic (PLEG): container finished" podID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerID="812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c" exitCode=143 Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.615150 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2e84d98e-c6f1-4551-9e45-7085e31720ae" containerName="nova-scheduler-scheduler" containerID="cri-o://1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9" gracePeriod=30 Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.615224 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d772af7-a796-4c1a-af7d-04c440a68d86","Type":"ContainerDied","Data":"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac"} Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.615244 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d772af7-a796-4c1a-af7d-04c440a68d86","Type":"ContainerDied","Data":"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c"} Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.615254 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d772af7-a796-4c1a-af7d-04c440a68d86","Type":"ContainerDied","Data":"bd2010c8a72dbd377d1494af9d10620cea6c34379728ce5931416383a43a93e5"} Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.615270 4675 scope.go:117] "RemoveContainer" containerID="5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.663865 4675 scope.go:117] "RemoveContainer" containerID="812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.714666 4675 scope.go:117] "RemoveContainer" containerID="5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac" Feb 19 09:03:52 crc kubenswrapper[4675]: E0219 09:03:52.715204 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac\": container with ID starting with 5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac not found: ID does not exist" containerID="5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.715255 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac"} err="failed to get container status \"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac\": rpc error: code = NotFound desc = could not find container \"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac\": container with ID starting with 5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac not found: ID does not exist" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.715283 4675 scope.go:117] "RemoveContainer" containerID="812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c" Feb 19 09:03:52 crc kubenswrapper[4675]: E0219 09:03:52.716241 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c\": container with ID starting with 812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c not found: ID does not exist" containerID="812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.716277 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c"} err="failed to get container status \"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c\": rpc error: code = NotFound desc = could not find container \"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c\": container with ID starting with 812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c not found: ID does not exist" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.716327 4675 scope.go:117] "RemoveContainer" containerID="5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.716596 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac"} err="failed to get container status \"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac\": rpc error: code = NotFound desc = could not find container \"5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac\": container with ID starting with 5faecb22707afa97a742a26d08f9e60853b78061a8bee310415cb45f684370ac not found: ID does not exist" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.716636 4675 scope.go:117] "RemoveContainer" containerID="812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.717136 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c"} err="failed to get container status \"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c\": rpc error: code = NotFound desc = could not find container \"812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c\": container with ID starting with 812dd677545a60468b0d7d8fab046f9f77955dfdd6df66ff1c88ee0081ad573c not found: ID does not exist" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.770840 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-combined-ca-bundle\") pod \"8d772af7-a796-4c1a-af7d-04c440a68d86\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.770960 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-nova-metadata-tls-certs\") pod \"8d772af7-a796-4c1a-af7d-04c440a68d86\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.771012 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vdmg\" (UniqueName: \"kubernetes.io/projected/8d772af7-a796-4c1a-af7d-04c440a68d86-kube-api-access-2vdmg\") pod \"8d772af7-a796-4c1a-af7d-04c440a68d86\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.771103 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-config-data\") pod \"8d772af7-a796-4c1a-af7d-04c440a68d86\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.773422 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d772af7-a796-4c1a-af7d-04c440a68d86-logs\") pod \"8d772af7-a796-4c1a-af7d-04c440a68d86\" (UID: \"8d772af7-a796-4c1a-af7d-04c440a68d86\") " Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.773492 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d772af7-a796-4c1a-af7d-04c440a68d86-logs" (OuterVolumeSpecName: "logs") pod "8d772af7-a796-4c1a-af7d-04c440a68d86" (UID: "8d772af7-a796-4c1a-af7d-04c440a68d86"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.774420 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d772af7-a796-4c1a-af7d-04c440a68d86-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.776452 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d772af7-a796-4c1a-af7d-04c440a68d86-kube-api-access-2vdmg" (OuterVolumeSpecName: "kube-api-access-2vdmg") pod "8d772af7-a796-4c1a-af7d-04c440a68d86" (UID: "8d772af7-a796-4c1a-af7d-04c440a68d86"). InnerVolumeSpecName "kube-api-access-2vdmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.799591 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d772af7-a796-4c1a-af7d-04c440a68d86" (UID: "8d772af7-a796-4c1a-af7d-04c440a68d86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.800118 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-config-data" (OuterVolumeSpecName: "config-data") pod "8d772af7-a796-4c1a-af7d-04c440a68d86" (UID: "8d772af7-a796-4c1a-af7d-04c440a68d86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.830608 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8d772af7-a796-4c1a-af7d-04c440a68d86" (UID: "8d772af7-a796-4c1a-af7d-04c440a68d86"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.877034 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.877075 4675 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.877092 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vdmg\" (UniqueName: \"kubernetes.io/projected/8d772af7-a796-4c1a-af7d-04c440a68d86-kube-api-access-2vdmg\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:52 crc kubenswrapper[4675]: I0219 09:03:52.877105 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d772af7-a796-4c1a-af7d-04c440a68d86-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.629740 4675 generic.go:334] "Generic (PLEG): container finished" podID="3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" containerID="3e3ce76b3c32778f16f95ff346231639cc54324dd25df20f79a0f043e8bd655a" exitCode=0 Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.629966 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" event={"ID":"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102","Type":"ContainerDied","Data":"3e3ce76b3c32778f16f95ff346231639cc54324dd25df20f79a0f043e8bd655a"} Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.638330 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.757086 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.777049 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.789268 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:53 crc kubenswrapper[4675]: E0219 09:03:53.789822 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-log" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.789846 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-log" Feb 19 09:03:53 crc kubenswrapper[4675]: E0219 09:03:53.789889 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5077fc1f-d1da-4683-84e8-52afdefcd0e1" containerName="nova-manage" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.789904 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="5077fc1f-d1da-4683-84e8-52afdefcd0e1" containerName="nova-manage" Feb 19 09:03:53 crc kubenswrapper[4675]: E0219 09:03:53.789930 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-metadata" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.789937 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-metadata" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.790136 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-log" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.790157 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="5077fc1f-d1da-4683-84e8-52afdefcd0e1" containerName="nova-manage" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.790182 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" containerName="nova-metadata-metadata" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.791460 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.799854 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.808167 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.808425 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.896115 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwr6g\" (UniqueName: \"kubernetes.io/projected/322bc0cb-2559-4c88-93a1-01a3cc10c29c-kube-api-access-cwr6g\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.896197 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.896227 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-config-data\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.896258 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322bc0cb-2559-4c88-93a1-01a3cc10c29c-logs\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.896278 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.997687 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwr6g\" (UniqueName: \"kubernetes.io/projected/322bc0cb-2559-4c88-93a1-01a3cc10c29c-kube-api-access-cwr6g\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.997751 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.997783 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-config-data\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.997816 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322bc0cb-2559-4c88-93a1-01a3cc10c29c-logs\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:53 crc kubenswrapper[4675]: I0219 09:03:53.997835 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:53.999129 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322bc0cb-2559-4c88-93a1-01a3cc10c29c-logs\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.011047 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.011219 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-config-data\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.024172 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.026299 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwr6g\" (UniqueName: \"kubernetes.io/projected/322bc0cb-2559-4c88-93a1-01a3cc10c29c-kube-api-access-cwr6g\") pod \"nova-metadata-0\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " pod="openstack/nova-metadata-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.117355 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.117842 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.201862 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jk5f\" (UniqueName: \"kubernetes.io/projected/2e84d98e-c6f1-4551-9e45-7085e31720ae-kube-api-access-7jk5f\") pod \"2e84d98e-c6f1-4551-9e45-7085e31720ae\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.202205 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-combined-ca-bundle\") pod \"2e84d98e-c6f1-4551-9e45-7085e31720ae\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.203295 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-config-data\") pod \"2e84d98e-c6f1-4551-9e45-7085e31720ae\" (UID: \"2e84d98e-c6f1-4551-9e45-7085e31720ae\") " Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.208849 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e84d98e-c6f1-4551-9e45-7085e31720ae-kube-api-access-7jk5f" (OuterVolumeSpecName: "kube-api-access-7jk5f") pod "2e84d98e-c6f1-4551-9e45-7085e31720ae" (UID: "2e84d98e-c6f1-4551-9e45-7085e31720ae"). InnerVolumeSpecName "kube-api-access-7jk5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.238766 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-config-data" (OuterVolumeSpecName: "config-data") pod "2e84d98e-c6f1-4551-9e45-7085e31720ae" (UID: "2e84d98e-c6f1-4551-9e45-7085e31720ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.261756 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e84d98e-c6f1-4551-9e45-7085e31720ae" (UID: "2e84d98e-c6f1-4551-9e45-7085e31720ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.307583 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jk5f\" (UniqueName: \"kubernetes.io/projected/2e84d98e-c6f1-4551-9e45-7085e31720ae-kube-api-access-7jk5f\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.307626 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.307649 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e84d98e-c6f1-4551-9e45-7085e31720ae-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.616087 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.647808 4675 generic.go:334] "Generic (PLEG): container finished" podID="2e84d98e-c6f1-4551-9e45-7085e31720ae" containerID="1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9" exitCode=0 Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.647867 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e84d98e-c6f1-4551-9e45-7085e31720ae","Type":"ContainerDied","Data":"1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9"} Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.647893 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2e84d98e-c6f1-4551-9e45-7085e31720ae","Type":"ContainerDied","Data":"0baaa3919cef179f32913aee966ba3dc6ca43583884b1d3a9837b33e35916047"} Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.647910 4675 scope.go:117] "RemoveContainer" containerID="1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.648024 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.650777 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"322bc0cb-2559-4c88-93a1-01a3cc10c29c","Type":"ContainerStarted","Data":"83aa733e217322b62971bab6c293bd91522afec07b9aea215bd976f4f5b3d325"} Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.683885 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.689974 4675 scope.go:117] "RemoveContainer" containerID="1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9" Feb 19 09:03:54 crc kubenswrapper[4675]: E0219 09:03:54.690323 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9\": container with ID starting with 1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9 not found: ID does not exist" containerID="1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.690361 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9"} err="failed to get container status \"1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9\": rpc error: code = NotFound desc = could not find container \"1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9\": container with ID starting with 1ab8a43d4ba297e22a9feebba24ed94015a64196d04473a783e0e8972f7de7e9 not found: ID does not exist" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.692213 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.708684 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:54 crc kubenswrapper[4675]: E0219 09:03:54.709558 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e84d98e-c6f1-4551-9e45-7085e31720ae" containerName="nova-scheduler-scheduler" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.709579 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e84d98e-c6f1-4551-9e45-7085e31720ae" containerName="nova-scheduler-scheduler" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.709774 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e84d98e-c6f1-4551-9e45-7085e31720ae" containerName="nova-scheduler-scheduler" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.710375 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.713823 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.726258 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.817390 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-config-data\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.817465 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2frw8\" (UniqueName: \"kubernetes.io/projected/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-kube-api-access-2frw8\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.817767 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.919915 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-config-data\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.919965 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2frw8\" (UniqueName: \"kubernetes.io/projected/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-kube-api-access-2frw8\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.920019 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.921262 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.925663 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-config-data\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.926131 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:54 crc kubenswrapper[4675]: I0219 09:03:54.941297 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2frw8\" (UniqueName: \"kubernetes.io/projected/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-kube-api-access-2frw8\") pod \"nova-scheduler-0\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " pod="openstack/nova-scheduler-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.021884 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data\") pod \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.022171 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-scripts\") pod \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.022272 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-combined-ca-bundle\") pod \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.022356 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6smfg\" (UniqueName: \"kubernetes.io/projected/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-kube-api-access-6smfg\") pod \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.026242 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-scripts" (OuterVolumeSpecName: "scripts") pod "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" (UID: "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.026885 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-kube-api-access-6smfg" (OuterVolumeSpecName: "kube-api-access-6smfg") pod "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" (UID: "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102"). InnerVolumeSpecName "kube-api-access-6smfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.029303 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:03:55 crc kubenswrapper[4675]: E0219 09:03:55.045850 4675 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data podName:3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102 nodeName:}" failed. No retries permitted until 2026-02-19 09:03:55.545814171 +0000 UTC m=+1217.172904439 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data") pod "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" (UID: "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102") : error deleting /var/lib/kubelet/pods/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102/volume-subpaths: remove /var/lib/kubelet/pods/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102/volume-subpaths: no such file or directory Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.050137 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" (UID: "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.068890 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.117819 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e84d98e-c6f1-4551-9e45-7085e31720ae" path="/var/lib/kubelet/pods/2e84d98e-c6f1-4551-9e45-7085e31720ae/volumes" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.118423 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d772af7-a796-4c1a-af7d-04c440a68d86" path="/var/lib/kubelet/pods/8d772af7-a796-4c1a-af7d-04c440a68d86/volumes" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.124704 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.124739 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.124749 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6smfg\" (UniqueName: \"kubernetes.io/projected/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-kube-api-access-6smfg\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.163149 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qr6ht"] Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.163429 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" podUID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerName="dnsmasq-dns" containerID="cri-o://1697cb8de82ab9deb11ca0300483af6c6ebaf528c52f0b463e5ba5235551636c" gracePeriod=10 Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.538385 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.636147 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data\") pod \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\" (UID: \"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.640827 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data" (OuterVolumeSpecName: "config-data") pod "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" (UID: "3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.673576 4675 generic.go:334] "Generic (PLEG): container finished" podID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerID="1697cb8de82ab9deb11ca0300483af6c6ebaf528c52f0b463e5ba5235551636c" exitCode=0 Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.673660 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" event={"ID":"825f8331-17ce-4e8e-beb5-4f2807fc4ceb","Type":"ContainerDied","Data":"1697cb8de82ab9deb11ca0300483af6c6ebaf528c52f0b463e5ba5235551636c"} Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.675719 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.676089 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8lmq6" event={"ID":"3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102","Type":"ContainerDied","Data":"9659fa92154b85665be4657d357ae5bf908ba64f59325b0f10275ef162156620"} Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.676127 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9659fa92154b85665be4657d357ae5bf908ba64f59325b0f10275ef162156620" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.682964 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a3bb7b0-46ad-488d-b4c1-7e486d09866f","Type":"ContainerStarted","Data":"0dcb7fd3db27d8ab37ffd08cee841351c6bbe7d05a7b5e73bf7ebd31d3d72258"} Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.685858 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.685901 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"322bc0cb-2559-4c88-93a1-01a3cc10c29c","Type":"ContainerStarted","Data":"f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f"} Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.685953 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"322bc0cb-2559-4c88-93a1-01a3cc10c29c","Type":"ContainerStarted","Data":"cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4"} Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.709682 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.709665659 podStartE2EDuration="2.709665659s" podCreationTimestamp="2026-02-19 09:03:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:55.708530599 +0000 UTC m=+1217.335620857" watchObservedRunningTime="2026-02-19 09:03:55.709665659 +0000 UTC m=+1217.336755927" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.737491 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44v27\" (UniqueName: \"kubernetes.io/projected/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-kube-api-access-44v27\") pod \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.737563 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-config\") pod \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.737680 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-swift-storage-0\") pod \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.737750 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-sb\") pod \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.737777 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-nb\") pod \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.737851 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-svc\") pod \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\" (UID: \"825f8331-17ce-4e8e-beb5-4f2807fc4ceb\") " Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.738392 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.751399 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-kube-api-access-44v27" (OuterVolumeSpecName: "kube-api-access-44v27") pod "825f8331-17ce-4e8e-beb5-4f2807fc4ceb" (UID: "825f8331-17ce-4e8e-beb5-4f2807fc4ceb"). InnerVolumeSpecName "kube-api-access-44v27". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.771237 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 19 09:03:55 crc kubenswrapper[4675]: E0219 09:03:55.772096 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerName="dnsmasq-dns" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.772118 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerName="dnsmasq-dns" Feb 19 09:03:55 crc kubenswrapper[4675]: E0219 09:03:55.772137 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" containerName="nova-cell1-conductor-db-sync" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.772196 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" containerName="nova-cell1-conductor-db-sync" Feb 19 09:03:55 crc kubenswrapper[4675]: E0219 09:03:55.772216 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerName="init" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.772225 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerName="init" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.772527 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" containerName="nova-cell1-conductor-db-sync" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.772556 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" containerName="dnsmasq-dns" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.774167 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.778287 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.790426 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.840108 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670df879-c26a-46f0-b385-5bb153734e18-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.840199 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670df879-c26a-46f0-b385-5bb153734e18-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.840262 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sqfx\" (UniqueName: \"kubernetes.io/projected/670df879-c26a-46f0-b385-5bb153734e18-kube-api-access-9sqfx\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.840344 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44v27\" (UniqueName: \"kubernetes.io/projected/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-kube-api-access-44v27\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.873737 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-config" (OuterVolumeSpecName: "config") pod "825f8331-17ce-4e8e-beb5-4f2807fc4ceb" (UID: "825f8331-17ce-4e8e-beb5-4f2807fc4ceb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.878039 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "825f8331-17ce-4e8e-beb5-4f2807fc4ceb" (UID: "825f8331-17ce-4e8e-beb5-4f2807fc4ceb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.879994 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "825f8331-17ce-4e8e-beb5-4f2807fc4ceb" (UID: "825f8331-17ce-4e8e-beb5-4f2807fc4ceb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.887614 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "825f8331-17ce-4e8e-beb5-4f2807fc4ceb" (UID: "825f8331-17ce-4e8e-beb5-4f2807fc4ceb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.893620 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "825f8331-17ce-4e8e-beb5-4f2807fc4ceb" (UID: "825f8331-17ce-4e8e-beb5-4f2807fc4ceb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942118 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670df879-c26a-46f0-b385-5bb153734e18-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942233 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670df879-c26a-46f0-b385-5bb153734e18-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942290 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sqfx\" (UniqueName: \"kubernetes.io/projected/670df879-c26a-46f0-b385-5bb153734e18-kube-api-access-9sqfx\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942382 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942398 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942411 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942421 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.942432 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/825f8331-17ce-4e8e-beb5-4f2807fc4ceb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.946273 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670df879-c26a-46f0-b385-5bb153734e18-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.947150 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670df879-c26a-46f0-b385-5bb153734e18-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:55 crc kubenswrapper[4675]: I0219 09:03:55.958769 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sqfx\" (UniqueName: \"kubernetes.io/projected/670df879-c26a-46f0-b385-5bb153734e18-kube-api-access-9sqfx\") pod \"nova-cell1-conductor-0\" (UID: \"670df879-c26a-46f0-b385-5bb153734e18\") " pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.120352 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.614765 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.679600 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.711905 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"670df879-c26a-46f0-b385-5bb153734e18","Type":"ContainerStarted","Data":"e790093b1f7060e50ae057a65572605c7f09236c8aac08cdd10aada02be2c6ba"} Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.717074 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" event={"ID":"825f8331-17ce-4e8e-beb5-4f2807fc4ceb","Type":"ContainerDied","Data":"f21a39d1faebc8547eb29ffb30a6ca416e32d59c7493d37d795c1a8a3f16a1c0"} Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.717148 4675 scope.go:117] "RemoveContainer" containerID="1697cb8de82ab9deb11ca0300483af6c6ebaf528c52f0b463e5ba5235551636c" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.717359 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-qr6ht" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.724120 4675 generic.go:334] "Generic (PLEG): container finished" podID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerID="300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79" exitCode=0 Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.724245 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8969cd9a-d085-42c8-910d-8cb09bbeb93b","Type":"ContainerDied","Data":"300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79"} Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.724275 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8969cd9a-d085-42c8-910d-8cb09bbeb93b","Type":"ContainerDied","Data":"965133f30b80317846b491770f4a04cb5119176eb5f1e5e1e433ecfa4107c685"} Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.724369 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.737108 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a3bb7b0-46ad-488d-b4c1-7e486d09866f","Type":"ContainerStarted","Data":"e2afacffb7ba83062ad36f29920716477e238540a693024edf571a4430ad8c05"} Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.761507 4675 scope.go:117] "RemoveContainer" containerID="5bd73ab1d57b0d778ca4fedc285167e3357cb43490ec9fc6d5ee18aa167b0dd2" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.781543 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7815219669999998 podStartE2EDuration="2.781521967s" podCreationTimestamp="2026-02-19 09:03:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:56.761154816 +0000 UTC m=+1218.388245084" watchObservedRunningTime="2026-02-19 09:03:56.781521967 +0000 UTC m=+1218.408612235" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.786296 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qr6ht"] Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.793699 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-qr6ht"] Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.821687 4675 scope.go:117] "RemoveContainer" containerID="300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.847067 4675 scope.go:117] "RemoveContainer" containerID="202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.867705 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjvrc\" (UniqueName: \"kubernetes.io/projected/8969cd9a-d085-42c8-910d-8cb09bbeb93b-kube-api-access-wjvrc\") pod \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.867864 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-config-data\") pod \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.867890 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-combined-ca-bundle\") pod \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.867952 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8969cd9a-d085-42c8-910d-8cb09bbeb93b-logs\") pod \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\" (UID: \"8969cd9a-d085-42c8-910d-8cb09bbeb93b\") " Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.872748 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8969cd9a-d085-42c8-910d-8cb09bbeb93b-logs" (OuterVolumeSpecName: "logs") pod "8969cd9a-d085-42c8-910d-8cb09bbeb93b" (UID: "8969cd9a-d085-42c8-910d-8cb09bbeb93b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.881930 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8969cd9a-d085-42c8-910d-8cb09bbeb93b-kube-api-access-wjvrc" (OuterVolumeSpecName: "kube-api-access-wjvrc") pod "8969cd9a-d085-42c8-910d-8cb09bbeb93b" (UID: "8969cd9a-d085-42c8-910d-8cb09bbeb93b"). InnerVolumeSpecName "kube-api-access-wjvrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.894921 4675 scope.go:117] "RemoveContainer" containerID="300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79" Feb 19 09:03:56 crc kubenswrapper[4675]: E0219 09:03:56.895405 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79\": container with ID starting with 300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79 not found: ID does not exist" containerID="300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.895457 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79"} err="failed to get container status \"300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79\": rpc error: code = NotFound desc = could not find container \"300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79\": container with ID starting with 300ab1a53956bc0b900ba0054b3a6db5aa7187d08a3d5fb0b1088d03a9d6de79 not found: ID does not exist" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.895477 4675 scope.go:117] "RemoveContainer" containerID="202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c" Feb 19 09:03:56 crc kubenswrapper[4675]: E0219 09:03:56.895807 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c\": container with ID starting with 202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c not found: ID does not exist" containerID="202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.895827 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c"} err="failed to get container status \"202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c\": rpc error: code = NotFound desc = could not find container \"202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c\": container with ID starting with 202b4b030eacbea7620246f01f967d6a1ba4ef412c91718dc77daad4785f604c not found: ID does not exist" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.897589 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8969cd9a-d085-42c8-910d-8cb09bbeb93b" (UID: "8969cd9a-d085-42c8-910d-8cb09bbeb93b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.904873 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-config-data" (OuterVolumeSpecName: "config-data") pod "8969cd9a-d085-42c8-910d-8cb09bbeb93b" (UID: "8969cd9a-d085-42c8-910d-8cb09bbeb93b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.969382 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8969cd9a-d085-42c8-910d-8cb09bbeb93b-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.969427 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjvrc\" (UniqueName: \"kubernetes.io/projected/8969cd9a-d085-42c8-910d-8cb09bbeb93b-kube-api-access-wjvrc\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.969439 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:56 crc kubenswrapper[4675]: I0219 09:03:56.969449 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969cd9a-d085-42c8-910d-8cb09bbeb93b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.097734 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.119407 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="825f8331-17ce-4e8e-beb5-4f2807fc4ceb" path="/var/lib/kubelet/pods/825f8331-17ce-4e8e-beb5-4f2807fc4ceb/volumes" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.120208 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.123550 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:57 crc kubenswrapper[4675]: E0219 09:03:57.123959 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-api" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.123982 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-api" Feb 19 09:03:57 crc kubenswrapper[4675]: E0219 09:03:57.124001 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-log" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.124008 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-log" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.124192 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-api" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.124214 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" containerName="nova-api-log" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.125219 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.129702 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.144891 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.281978 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-config-data\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.282055 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.282107 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73abfd4a-24a9-4dc1-8064-1297d004af92-logs\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.282171 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d7nm\" (UniqueName: \"kubernetes.io/projected/73abfd4a-24a9-4dc1-8064-1297d004af92-kube-api-access-8d7nm\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.383691 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d7nm\" (UniqueName: \"kubernetes.io/projected/73abfd4a-24a9-4dc1-8064-1297d004af92-kube-api-access-8d7nm\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.383859 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-config-data\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.383888 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.383909 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73abfd4a-24a9-4dc1-8064-1297d004af92-logs\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.384333 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73abfd4a-24a9-4dc1-8064-1297d004af92-logs\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.393345 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.393441 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-config-data\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.427190 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d7nm\" (UniqueName: \"kubernetes.io/projected/73abfd4a-24a9-4dc1-8064-1297d004af92-kube-api-access-8d7nm\") pod \"nova-api-0\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.449124 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.748214 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"670df879-c26a-46f0-b385-5bb153734e18","Type":"ContainerStarted","Data":"764ba4ae54dd5e35fb17284204e9e55f57889f79f64839a66ce0b6102d2be772"} Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.748808 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.771884 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.771864542 podStartE2EDuration="2.771864542s" podCreationTimestamp="2026-02-19 09:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:57.76183378 +0000 UTC m=+1219.388924048" watchObservedRunningTime="2026-02-19 09:03:57.771864542 +0000 UTC m=+1219.398954800" Feb 19 09:03:57 crc kubenswrapper[4675]: I0219 09:03:57.956858 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:03:57 crc kubenswrapper[4675]: W0219 09:03:57.959901 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73abfd4a_24a9_4dc1_8064_1297d004af92.slice/crio-e0ecdc6c35596297da403e9a4a0669ef35dd778ef76b09dea964d059a4a4e267 WatchSource:0}: Error finding container e0ecdc6c35596297da403e9a4a0669ef35dd778ef76b09dea964d059a4a4e267: Status 404 returned error can't find the container with id e0ecdc6c35596297da403e9a4a0669ef35dd778ef76b09dea964d059a4a4e267 Feb 19 09:03:58 crc kubenswrapper[4675]: I0219 09:03:58.763780 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73abfd4a-24a9-4dc1-8064-1297d004af92","Type":"ContainerStarted","Data":"cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c"} Feb 19 09:03:58 crc kubenswrapper[4675]: I0219 09:03:58.763827 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73abfd4a-24a9-4dc1-8064-1297d004af92","Type":"ContainerStarted","Data":"c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d"} Feb 19 09:03:58 crc kubenswrapper[4675]: I0219 09:03:58.763838 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73abfd4a-24a9-4dc1-8064-1297d004af92","Type":"ContainerStarted","Data":"e0ecdc6c35596297da403e9a4a0669ef35dd778ef76b09dea964d059a4a4e267"} Feb 19 09:03:58 crc kubenswrapper[4675]: I0219 09:03:58.790892 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.790867384 podStartE2EDuration="1.790867384s" podCreationTimestamp="2026-02-19 09:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:03:58.782585628 +0000 UTC m=+1220.409675916" watchObservedRunningTime="2026-02-19 09:03:58.790867384 +0000 UTC m=+1220.417957652" Feb 19 09:03:59 crc kubenswrapper[4675]: I0219 09:03:59.123851 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8969cd9a-d085-42c8-910d-8cb09bbeb93b" path="/var/lib/kubelet/pods/8969cd9a-d085-42c8-910d-8cb09bbeb93b/volumes" Feb 19 09:03:59 crc kubenswrapper[4675]: I0219 09:03:59.125182 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 19 09:03:59 crc kubenswrapper[4675]: I0219 09:03:59.125216 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 19 09:04:00 crc kubenswrapper[4675]: I0219 09:04:00.029720 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 19 09:04:01 crc kubenswrapper[4675]: I0219 09:04:01.152054 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 19 09:04:04 crc kubenswrapper[4675]: I0219 09:04:04.118374 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 19 09:04:04 crc kubenswrapper[4675]: I0219 09:04:04.118927 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 19 09:04:04 crc kubenswrapper[4675]: I0219 09:04:04.973703 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 19 09:04:05 crc kubenswrapper[4675]: I0219 09:04:05.030463 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 19 09:04:05 crc kubenswrapper[4675]: I0219 09:04:05.062283 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 19 09:04:05 crc kubenswrapper[4675]: I0219 09:04:05.136833 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:05 crc kubenswrapper[4675]: I0219 09:04:05.136852 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:05 crc kubenswrapper[4675]: I0219 09:04:05.851031 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 19 09:04:07 crc kubenswrapper[4675]: I0219 09:04:07.449747 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:04:07 crc kubenswrapper[4675]: I0219 09:04:07.450050 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:04:08 crc kubenswrapper[4675]: I0219 09:04:08.531994 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:08 crc kubenswrapper[4675]: I0219 09:04:08.532036 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:08 crc kubenswrapper[4675]: I0219 09:04:08.794477 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 09:04:08 crc kubenswrapper[4675]: I0219 09:04:08.794719 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="ba2e218a-edb2-4626-b480-1e4a534bc80d" containerName="kube-state-metrics" containerID="cri-o://7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3" gracePeriod=30 Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.361547 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.459981 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnhw4\" (UniqueName: \"kubernetes.io/projected/ba2e218a-edb2-4626-b480-1e4a534bc80d-kube-api-access-cnhw4\") pod \"ba2e218a-edb2-4626-b480-1e4a534bc80d\" (UID: \"ba2e218a-edb2-4626-b480-1e4a534bc80d\") " Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.480611 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba2e218a-edb2-4626-b480-1e4a534bc80d-kube-api-access-cnhw4" (OuterVolumeSpecName: "kube-api-access-cnhw4") pod "ba2e218a-edb2-4626-b480-1e4a534bc80d" (UID: "ba2e218a-edb2-4626-b480-1e4a534bc80d"). InnerVolumeSpecName "kube-api-access-cnhw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.562417 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnhw4\" (UniqueName: \"kubernetes.io/projected/ba2e218a-edb2-4626-b480-1e4a534bc80d-kube-api-access-cnhw4\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.860891 4675 generic.go:334] "Generic (PLEG): container finished" podID="ba2e218a-edb2-4626-b480-1e4a534bc80d" containerID="7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3" exitCode=2 Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.860947 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ba2e218a-edb2-4626-b480-1e4a534bc80d","Type":"ContainerDied","Data":"7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3"} Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.860975 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ba2e218a-edb2-4626-b480-1e4a534bc80d","Type":"ContainerDied","Data":"7b32ea5c053a87da0165bf77be36ad6a13bdbef9c109973347a38b206c0e99aa"} Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.860993 4675 scope.go:117] "RemoveContainer" containerID="7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.861017 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.897968 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.901034 4675 scope.go:117] "RemoveContainer" containerID="7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3" Feb 19 09:04:09 crc kubenswrapper[4675]: E0219 09:04:09.901658 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3\": container with ID starting with 7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3 not found: ID does not exist" containerID="7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.901703 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3"} err="failed to get container status \"7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3\": rpc error: code = NotFound desc = could not find container \"7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3\": container with ID starting with 7435616bb1123dd8027a2bcbbd0d48348c011fd941a7f5db4d03d718656f3ce3 not found: ID does not exist" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.918513 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.929052 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 09:04:09 crc kubenswrapper[4675]: E0219 09:04:09.929620 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2e218a-edb2-4626-b480-1e4a534bc80d" containerName="kube-state-metrics" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.929662 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2e218a-edb2-4626-b480-1e4a534bc80d" containerName="kube-state-metrics" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.929900 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba2e218a-edb2-4626-b480-1e4a534bc80d" containerName="kube-state-metrics" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.930691 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.932839 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.934447 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 19 09:04:09 crc kubenswrapper[4675]: I0219 09:04:09.937661 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.072202 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.072690 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.072784 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.072820 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twztj\" (UniqueName: \"kubernetes.io/projected/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-api-access-twztj\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.175120 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.175194 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twztj\" (UniqueName: \"kubernetes.io/projected/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-api-access-twztj\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.175326 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.175366 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.180584 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.180937 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.182921 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e70d121-b4ec-4c18-be34-18d2eb7d529e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.199208 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twztj\" (UniqueName: \"kubernetes.io/projected/1e70d121-b4ec-4c18-be34-18d2eb7d529e-kube-api-access-twztj\") pod \"kube-state-metrics-0\" (UID: \"1e70d121-b4ec-4c18-be34-18d2eb7d529e\") " pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.253046 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.722714 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 19 09:04:10 crc kubenswrapper[4675]: W0219 09:04:10.739657 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e70d121_b4ec_4c18_be34_18d2eb7d529e.slice/crio-c4bbe0fde83cad16d6528377180ffe1ff9922410cd5f1091e6b1b29de17b8869 WatchSource:0}: Error finding container c4bbe0fde83cad16d6528377180ffe1ff9922410cd5f1091e6b1b29de17b8869: Status 404 returned error can't find the container with id c4bbe0fde83cad16d6528377180ffe1ff9922410cd5f1091e6b1b29de17b8869 Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.872661 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e70d121-b4ec-4c18-be34-18d2eb7d529e","Type":"ContainerStarted","Data":"c4bbe0fde83cad16d6528377180ffe1ff9922410cd5f1091e6b1b29de17b8869"} Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.970200 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.970469 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-central-agent" containerID="cri-o://98caa5676c91647f6dcd5bd202334094f163141ac453c1f606241c1c3c04927f" gracePeriod=30 Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.970600 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="proxy-httpd" containerID="cri-o://9dac50a6e901cab039ecef2c07f3d1c043af5dccbc418fcecefedccc70f3ca98" gracePeriod=30 Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.970677 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="sg-core" containerID="cri-o://2ce9eddfa923ad9b9a7e17f16a704c8123f9ba55ae63b2505fd9dc9368bf307d" gracePeriod=30 Feb 19 09:04:10 crc kubenswrapper[4675]: I0219 09:04:10.970713 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-notification-agent" containerID="cri-o://8fd247fc8e94099dde6448180737030d5004728475f60b8fdde291f8835ec9f2" gracePeriod=30 Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.115428 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba2e218a-edb2-4626-b480-1e4a534bc80d" path="/var/lib/kubelet/pods/ba2e218a-edb2-4626-b480-1e4a534bc80d/volumes" Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.744048 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.744444 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.883658 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e70d121-b4ec-4c18-be34-18d2eb7d529e","Type":"ContainerStarted","Data":"2bf565c38a1166e3c091da18ee9f1f2bec1cc7aac9da75334d0befe49834921a"} Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.883815 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.887076 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerID="9dac50a6e901cab039ecef2c07f3d1c043af5dccbc418fcecefedccc70f3ca98" exitCode=0 Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.887108 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerID="2ce9eddfa923ad9b9a7e17f16a704c8123f9ba55ae63b2505fd9dc9368bf307d" exitCode=2 Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.887123 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerID="98caa5676c91647f6dcd5bd202334094f163141ac453c1f606241c1c3c04927f" exitCode=0 Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.887146 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerDied","Data":"9dac50a6e901cab039ecef2c07f3d1c043af5dccbc418fcecefedccc70f3ca98"} Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.887185 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerDied","Data":"2ce9eddfa923ad9b9a7e17f16a704c8123f9ba55ae63b2505fd9dc9368bf307d"} Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.887198 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerDied","Data":"98caa5676c91647f6dcd5bd202334094f163141ac453c1f606241c1c3c04927f"} Feb 19 09:04:11 crc kubenswrapper[4675]: I0219 09:04:11.903325 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.310517778 podStartE2EDuration="2.903303604s" podCreationTimestamp="2026-02-19 09:04:09 +0000 UTC" firstStartedPulling="2026-02-19 09:04:10.740084385 +0000 UTC m=+1232.367174653" lastFinishedPulling="2026-02-19 09:04:11.332870211 +0000 UTC m=+1232.959960479" observedRunningTime="2026-02-19 09:04:11.899481695 +0000 UTC m=+1233.526571963" watchObservedRunningTime="2026-02-19 09:04:11.903303604 +0000 UTC m=+1233.530393862" Feb 19 09:04:14 crc kubenswrapper[4675]: I0219 09:04:14.125112 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 19 09:04:14 crc kubenswrapper[4675]: I0219 09:04:14.125545 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 19 09:04:14 crc kubenswrapper[4675]: I0219 09:04:14.129714 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 19 09:04:14 crc kubenswrapper[4675]: I0219 09:04:14.130863 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 19 09:04:14 crc kubenswrapper[4675]: I0219 09:04:14.916462 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerID="8fd247fc8e94099dde6448180737030d5004728475f60b8fdde291f8835ec9f2" exitCode=0 Feb 19 09:04:14 crc kubenswrapper[4675]: I0219 09:04:14.917398 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerDied","Data":"8fd247fc8e94099dde6448180737030d5004728475f60b8fdde291f8835ec9f2"} Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.161160 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.175971 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-run-httpd\") pod \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176076 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65gfj\" (UniqueName: \"kubernetes.io/projected/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-kube-api-access-65gfj\") pod \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176155 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-log-httpd\") pod \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176226 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-sg-core-conf-yaml\") pod \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176249 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-scripts\") pod \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176334 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-combined-ca-bundle\") pod \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176356 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" (UID: "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176388 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-config-data\") pod \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\" (UID: \"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005\") " Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176583 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" (UID: "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176967 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.176992 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.182905 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-scripts" (OuterVolumeSpecName: "scripts") pod "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" (UID: "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.193994 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-kube-api-access-65gfj" (OuterVolumeSpecName: "kube-api-access-65gfj") pod "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" (UID: "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005"). InnerVolumeSpecName "kube-api-access-65gfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.219446 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" (UID: "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.279342 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.279377 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.279390 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65gfj\" (UniqueName: \"kubernetes.io/projected/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-kube-api-access-65gfj\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.281192 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" (UID: "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.301875 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-config-data" (OuterVolumeSpecName: "config-data") pod "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" (UID: "7a75e419-2bf4-4b7d-a5c4-f5f9a6135005"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.381006 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.381050 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.928858 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a75e419-2bf4-4b7d-a5c4-f5f9a6135005","Type":"ContainerDied","Data":"350a30b2c1263a9f0f532edc6181d31023c405138a6de80b79c7e4ef6cbb5f2e"} Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.928918 4675 scope.go:117] "RemoveContainer" containerID="9dac50a6e901cab039ecef2c07f3d1c043af5dccbc418fcecefedccc70f3ca98" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.928919 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.951697 4675 scope.go:117] "RemoveContainer" containerID="2ce9eddfa923ad9b9a7e17f16a704c8123f9ba55ae63b2505fd9dc9368bf307d" Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.969750 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:15 crc kubenswrapper[4675]: I0219 09:04:15.978796 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.001840 4675 scope.go:117] "RemoveContainer" containerID="8fd247fc8e94099dde6448180737030d5004728475f60b8fdde291f8835ec9f2" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.002657 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:16 crc kubenswrapper[4675]: E0219 09:04:16.003282 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-notification-agent" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003309 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-notification-agent" Feb 19 09:04:16 crc kubenswrapper[4675]: E0219 09:04:16.003336 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-central-agent" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003344 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-central-agent" Feb 19 09:04:16 crc kubenswrapper[4675]: E0219 09:04:16.003366 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="proxy-httpd" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003374 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="proxy-httpd" Feb 19 09:04:16 crc kubenswrapper[4675]: E0219 09:04:16.003413 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="sg-core" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003422 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="sg-core" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003661 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-central-agent" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003683 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="ceilometer-notification-agent" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003695 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="sg-core" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.003712 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" containerName="proxy-httpd" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.005853 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.008407 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.008685 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.009916 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.013280 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.042296 4675 scope.go:117] "RemoveContainer" containerID="98caa5676c91647f6dcd5bd202334094f163141ac453c1f606241c1c3c04927f" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.096505 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.096862 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-run-httpd\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.096973 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-log-httpd\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.097081 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-config-data\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.097300 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.097416 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r56v2\" (UniqueName: \"kubernetes.io/projected/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-kube-api-access-r56v2\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.097507 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-scripts\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.097585 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.199838 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.200099 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-run-httpd\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.200128 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-log-httpd\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.200153 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-config-data\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.200256 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.200324 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r56v2\" (UniqueName: \"kubernetes.io/projected/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-kube-api-access-r56v2\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.200362 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-scripts\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.200385 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.201403 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-log-httpd\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.202359 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-run-httpd\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.204937 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.206488 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-scripts\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.207500 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-config-data\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.207659 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.224208 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.225178 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r56v2\" (UniqueName: \"kubernetes.io/projected/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-kube-api-access-r56v2\") pod \"ceilometer-0\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.329881 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.767859 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:16 crc kubenswrapper[4675]: W0219 09:04:16.777196 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf73614d7_c8d1_40bd_93dc_0bf5dcd82624.slice/crio-7b50301d5789a09faebca309110c13736879204e0a23f8e297086c6df2d0c726 WatchSource:0}: Error finding container 7b50301d5789a09faebca309110c13736879204e0a23f8e297086c6df2d0c726: Status 404 returned error can't find the container with id 7b50301d5789a09faebca309110c13736879204e0a23f8e297086c6df2d0c726 Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.910371 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.938482 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerStarted","Data":"7b50301d5789a09faebca309110c13736879204e0a23f8e297086c6df2d0c726"} Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.941267 4675 generic.go:334] "Generic (PLEG): container finished" podID="15194877-7f24-46c5-aab3-229f004a3b7d" containerID="8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9" exitCode=137 Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.941318 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15194877-7f24-46c5-aab3-229f004a3b7d","Type":"ContainerDied","Data":"8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9"} Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.941350 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15194877-7f24-46c5-aab3-229f004a3b7d","Type":"ContainerDied","Data":"bb9a05df624375bf0278d6c9b8382ee8d7ab4f5f2a68c571fc360d98264ad1f8"} Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.941367 4675 scope.go:117] "RemoveContainer" containerID="8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.941470 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.968077 4675 scope.go:117] "RemoveContainer" containerID="8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9" Feb 19 09:04:16 crc kubenswrapper[4675]: E0219 09:04:16.968865 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9\": container with ID starting with 8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9 not found: ID does not exist" containerID="8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9" Feb 19 09:04:16 crc kubenswrapper[4675]: I0219 09:04:16.968920 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9"} err="failed to get container status \"8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9\": rpc error: code = NotFound desc = could not find container \"8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9\": container with ID starting with 8aefa42cd0bad56917593bb1addbd45e0bbb9031f00ff98c28f0e50a932dcfe9 not found: ID does not exist" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.022500 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-config-data\") pod \"15194877-7f24-46c5-aab3-229f004a3b7d\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.022583 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-combined-ca-bundle\") pod \"15194877-7f24-46c5-aab3-229f004a3b7d\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.022713 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bswpj\" (UniqueName: \"kubernetes.io/projected/15194877-7f24-46c5-aab3-229f004a3b7d-kube-api-access-bswpj\") pod \"15194877-7f24-46c5-aab3-229f004a3b7d\" (UID: \"15194877-7f24-46c5-aab3-229f004a3b7d\") " Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.029221 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15194877-7f24-46c5-aab3-229f004a3b7d-kube-api-access-bswpj" (OuterVolumeSpecName: "kube-api-access-bswpj") pod "15194877-7f24-46c5-aab3-229f004a3b7d" (UID: "15194877-7f24-46c5-aab3-229f004a3b7d"). InnerVolumeSpecName "kube-api-access-bswpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.051628 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-config-data" (OuterVolumeSpecName: "config-data") pod "15194877-7f24-46c5-aab3-229f004a3b7d" (UID: "15194877-7f24-46c5-aab3-229f004a3b7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.052589 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15194877-7f24-46c5-aab3-229f004a3b7d" (UID: "15194877-7f24-46c5-aab3-229f004a3b7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.114230 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a75e419-2bf4-4b7d-a5c4-f5f9a6135005" path="/var/lib/kubelet/pods/7a75e419-2bf4-4b7d-a5c4-f5f9a6135005/volumes" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.125720 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bswpj\" (UniqueName: \"kubernetes.io/projected/15194877-7f24-46c5-aab3-229f004a3b7d-kube-api-access-bswpj\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.125757 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.125772 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15194877-7f24-46c5-aab3-229f004a3b7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.262079 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.270243 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.284983 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:04:17 crc kubenswrapper[4675]: E0219 09:04:17.285462 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15194877-7f24-46c5-aab3-229f004a3b7d" containerName="nova-cell1-novncproxy-novncproxy" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.285489 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="15194877-7f24-46c5-aab3-229f004a3b7d" containerName="nova-cell1-novncproxy-novncproxy" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.285746 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="15194877-7f24-46c5-aab3-229f004a3b7d" containerName="nova-cell1-novncproxy-novncproxy" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.286490 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.296716 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.296924 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.298993 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.301942 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.328443 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgg2s\" (UniqueName: \"kubernetes.io/projected/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-kube-api-access-mgg2s\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.328591 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.328630 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.328668 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.328831 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.431273 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.431444 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgg2s\" (UniqueName: \"kubernetes.io/projected/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-kube-api-access-mgg2s\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.431548 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.431580 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.431598 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.435762 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.436388 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.436585 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.437897 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.451948 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgg2s\" (UniqueName: \"kubernetes.io/projected/774eaf0e-fab7-4e49-b3de-b9e8ca212f66-kube-api-access-mgg2s\") pod \"nova-cell1-novncproxy-0\" (UID: \"774eaf0e-fab7-4e49-b3de-b9e8ca212f66\") " pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.453858 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.454390 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.455025 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.459280 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.604676 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.957595 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerStarted","Data":"430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4"} Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.957922 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 19 09:04:17 crc kubenswrapper[4675]: I0219 09:04:17.962804 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.103507 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.172411 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-bn56v"] Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.174117 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.191144 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-bn56v"] Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.251851 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.251984 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-config\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.252120 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.252241 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.252366 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzscp\" (UniqueName: \"kubernetes.io/projected/634e48c1-5036-4a11-863d-b66f1c4fb341-kube-api-access-qzscp\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.252603 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.353918 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzscp\" (UniqueName: \"kubernetes.io/projected/634e48c1-5036-4a11-863d-b66f1c4fb341-kube-api-access-qzscp\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.353988 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.354075 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.354113 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-config\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.354138 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.354162 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.356351 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.357259 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.357449 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-config\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.357535 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.358364 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.374264 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzscp\" (UniqueName: \"kubernetes.io/projected/634e48c1-5036-4a11-863d-b66f1c4fb341-kube-api-access-qzscp\") pod \"dnsmasq-dns-5c7b6c5df9-bn56v\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.659697 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.984713 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"774eaf0e-fab7-4e49-b3de-b9e8ca212f66","Type":"ContainerStarted","Data":"e530d6aacd59963b50afe933fa3777d34dd22a3e862f0f88cb8166164d5a449b"} Feb 19 09:04:18 crc kubenswrapper[4675]: I0219 09:04:18.985041 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"774eaf0e-fab7-4e49-b3de-b9e8ca212f66","Type":"ContainerStarted","Data":"814d8e3139cf35a5a1c17197343206b9d7a7d1c3180bb4150d115be157b3b80a"} Feb 19 09:04:19 crc kubenswrapper[4675]: I0219 09:04:19.011260 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.011232331 podStartE2EDuration="2.011232331s" podCreationTimestamp="2026-02-19 09:04:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:04:19.002992606 +0000 UTC m=+1240.630082874" watchObservedRunningTime="2026-02-19 09:04:19.011232331 +0000 UTC m=+1240.638322599" Feb 19 09:04:19 crc kubenswrapper[4675]: I0219 09:04:19.139244 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15194877-7f24-46c5-aab3-229f004a3b7d" path="/var/lib/kubelet/pods/15194877-7f24-46c5-aab3-229f004a3b7d/volumes" Feb 19 09:04:19 crc kubenswrapper[4675]: I0219 09:04:19.334598 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-bn56v"] Feb 19 09:04:19 crc kubenswrapper[4675]: I0219 09:04:19.995844 4675 generic.go:334] "Generic (PLEG): container finished" podID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerID="77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a" exitCode=0 Feb 19 09:04:19 crc kubenswrapper[4675]: I0219 09:04:19.996690 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" event={"ID":"634e48c1-5036-4a11-863d-b66f1c4fb341","Type":"ContainerDied","Data":"77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a"} Feb 19 09:04:19 crc kubenswrapper[4675]: I0219 09:04:19.996750 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" event={"ID":"634e48c1-5036-4a11-863d-b66f1c4fb341","Type":"ContainerStarted","Data":"135f13e11f7e61b51ada72a1180847e5aeb36a92849fbb7b255fec087b19113a"} Feb 19 09:04:20 crc kubenswrapper[4675]: I0219 09:04:20.009440 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerStarted","Data":"1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d"} Feb 19 09:04:20 crc kubenswrapper[4675]: I0219 09:04:20.009753 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerStarted","Data":"61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4"} Feb 19 09:04:20 crc kubenswrapper[4675]: I0219 09:04:20.315028 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 19 09:04:21 crc kubenswrapper[4675]: I0219 09:04:21.019152 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" event={"ID":"634e48c1-5036-4a11-863d-b66f1c4fb341","Type":"ContainerStarted","Data":"89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d"} Feb 19 09:04:21 crc kubenswrapper[4675]: I0219 09:04:21.020832 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:21 crc kubenswrapper[4675]: I0219 09:04:21.050541 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" podStartSLOduration=3.050515208 podStartE2EDuration="3.050515208s" podCreationTimestamp="2026-02-19 09:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:04:21.045278301 +0000 UTC m=+1242.672368579" watchObservedRunningTime="2026-02-19 09:04:21.050515208 +0000 UTC m=+1242.677605466" Feb 19 09:04:21 crc kubenswrapper[4675]: I0219 09:04:21.281058 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:21 crc kubenswrapper[4675]: I0219 09:04:21.281628 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-log" containerID="cri-o://c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d" gracePeriod=30 Feb 19 09:04:21 crc kubenswrapper[4675]: I0219 09:04:21.281696 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-api" containerID="cri-o://cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c" gracePeriod=30 Feb 19 09:04:22 crc kubenswrapper[4675]: I0219 09:04:22.039283 4675 generic.go:334] "Generic (PLEG): container finished" podID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerID="c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d" exitCode=143 Feb 19 09:04:22 crc kubenswrapper[4675]: I0219 09:04:22.039347 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73abfd4a-24a9-4dc1-8064-1297d004af92","Type":"ContainerDied","Data":"c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d"} Feb 19 09:04:22 crc kubenswrapper[4675]: I0219 09:04:22.451510 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:22 crc kubenswrapper[4675]: I0219 09:04:22.605495 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:23 crc kubenswrapper[4675]: I0219 09:04:23.052275 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerStarted","Data":"094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4"} Feb 19 09:04:23 crc kubenswrapper[4675]: I0219 09:04:23.052352 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 19 09:04:23 crc kubenswrapper[4675]: I0219 09:04:23.076976 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9627091 podStartE2EDuration="8.076953048s" podCreationTimestamp="2026-02-19 09:04:15 +0000 UTC" firstStartedPulling="2026-02-19 09:04:16.786748829 +0000 UTC m=+1238.413839097" lastFinishedPulling="2026-02-19 09:04:21.900992777 +0000 UTC m=+1243.528083045" observedRunningTime="2026-02-19 09:04:23.071523207 +0000 UTC m=+1244.698613475" watchObservedRunningTime="2026-02-19 09:04:23.076953048 +0000 UTC m=+1244.704043316" Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.059433 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-central-agent" containerID="cri-o://430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4" gracePeriod=30 Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.059788 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="proxy-httpd" containerID="cri-o://094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4" gracePeriod=30 Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.059808 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="sg-core" containerID="cri-o://1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d" gracePeriod=30 Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.059817 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-notification-agent" containerID="cri-o://61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4" gracePeriod=30 Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.816964 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.991292 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-config-data\") pod \"73abfd4a-24a9-4dc1-8064-1297d004af92\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.991441 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-combined-ca-bundle\") pod \"73abfd4a-24a9-4dc1-8064-1297d004af92\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.991542 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73abfd4a-24a9-4dc1-8064-1297d004af92-logs\") pod \"73abfd4a-24a9-4dc1-8064-1297d004af92\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.991684 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d7nm\" (UniqueName: \"kubernetes.io/projected/73abfd4a-24a9-4dc1-8064-1297d004af92-kube-api-access-8d7nm\") pod \"73abfd4a-24a9-4dc1-8064-1297d004af92\" (UID: \"73abfd4a-24a9-4dc1-8064-1297d004af92\") " Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.992379 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73abfd4a-24a9-4dc1-8064-1297d004af92-logs" (OuterVolumeSpecName: "logs") pod "73abfd4a-24a9-4dc1-8064-1297d004af92" (UID: "73abfd4a-24a9-4dc1-8064-1297d004af92"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:04:24 crc kubenswrapper[4675]: I0219 09:04:24.997271 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73abfd4a-24a9-4dc1-8064-1297d004af92-kube-api-access-8d7nm" (OuterVolumeSpecName: "kube-api-access-8d7nm") pod "73abfd4a-24a9-4dc1-8064-1297d004af92" (UID: "73abfd4a-24a9-4dc1-8064-1297d004af92"). InnerVolumeSpecName "kube-api-access-8d7nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.024223 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-config-data" (OuterVolumeSpecName: "config-data") pod "73abfd4a-24a9-4dc1-8064-1297d004af92" (UID: "73abfd4a-24a9-4dc1-8064-1297d004af92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.032997 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73abfd4a-24a9-4dc1-8064-1297d004af92" (UID: "73abfd4a-24a9-4dc1-8064-1297d004af92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.069979 4675 generic.go:334] "Generic (PLEG): container finished" podID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerID="cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c" exitCode=0 Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.070052 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.070059 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73abfd4a-24a9-4dc1-8064-1297d004af92","Type":"ContainerDied","Data":"cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c"} Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.070131 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73abfd4a-24a9-4dc1-8064-1297d004af92","Type":"ContainerDied","Data":"e0ecdc6c35596297da403e9a4a0669ef35dd778ef76b09dea964d059a4a4e267"} Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.070151 4675 scope.go:117] "RemoveContainer" containerID="cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.074724 4675 generic.go:334] "Generic (PLEG): container finished" podID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerID="094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4" exitCode=0 Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.074750 4675 generic.go:334] "Generic (PLEG): container finished" podID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerID="1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d" exitCode=2 Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.074758 4675 generic.go:334] "Generic (PLEG): container finished" podID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerID="61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4" exitCode=0 Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.074756 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerDied","Data":"094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4"} Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.074797 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerDied","Data":"1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d"} Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.074813 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerDied","Data":"61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4"} Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.094738 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.094783 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73abfd4a-24a9-4dc1-8064-1297d004af92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.094799 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73abfd4a-24a9-4dc1-8064-1297d004af92-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.094810 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d7nm\" (UniqueName: \"kubernetes.io/projected/73abfd4a-24a9-4dc1-8064-1297d004af92-kube-api-access-8d7nm\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.096526 4675 scope.go:117] "RemoveContainer" containerID="c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.113192 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.130494 4675 scope.go:117] "RemoveContainer" containerID="cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c" Feb 19 09:04:25 crc kubenswrapper[4675]: E0219 09:04:25.131185 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c\": container with ID starting with cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c not found: ID does not exist" containerID="cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.131242 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c"} err="failed to get container status \"cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c\": rpc error: code = NotFound desc = could not find container \"cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c\": container with ID starting with cc6cdd953adb949b866a12fe807ae89b2ad065abf532a811e70daa0130e7eb4c not found: ID does not exist" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.131272 4675 scope.go:117] "RemoveContainer" containerID="c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d" Feb 19 09:04:25 crc kubenswrapper[4675]: E0219 09:04:25.131964 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d\": container with ID starting with c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d not found: ID does not exist" containerID="c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.131989 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d"} err="failed to get container status \"c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d\": rpc error: code = NotFound desc = could not find container \"c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d\": container with ID starting with c60d9c06a839fd8ac847d49db9dbc89f20885c97d8484a3490a4a8e76bfdc29d not found: ID does not exist" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.134142 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.176093 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:25 crc kubenswrapper[4675]: E0219 09:04:25.176646 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-log" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.176683 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-log" Feb 19 09:04:25 crc kubenswrapper[4675]: E0219 09:04:25.176708 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-api" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.176719 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-api" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.176931 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-api" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.176951 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" containerName="nova-api-log" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.178185 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.181091 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.181371 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.181514 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.204242 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.305613 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c28f0f-c32d-4114-b61e-592c108080e2-logs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.305743 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-public-tls-certs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.305770 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.305791 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.305956 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-config-data\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.305992 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xhhj\" (UniqueName: \"kubernetes.io/projected/78c28f0f-c32d-4114-b61e-592c108080e2-kube-api-access-9xhhj\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.408167 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-config-data\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.408225 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xhhj\" (UniqueName: \"kubernetes.io/projected/78c28f0f-c32d-4114-b61e-592c108080e2-kube-api-access-9xhhj\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.408273 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c28f0f-c32d-4114-b61e-592c108080e2-logs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.408312 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-public-tls-certs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.408330 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.408345 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.409302 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c28f0f-c32d-4114-b61e-592c108080e2-logs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.414509 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.414559 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-config-data\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.414809 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.415564 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-public-tls-certs\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.432566 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xhhj\" (UniqueName: \"kubernetes.io/projected/78c28f0f-c32d-4114-b61e-592c108080e2-kube-api-access-9xhhj\") pod \"nova-api-0\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.555365 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.669556 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.820222 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-run-httpd\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.820809 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-scripts\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.820805 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.821030 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-sg-core-conf-yaml\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.821116 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-combined-ca-bundle\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.821148 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r56v2\" (UniqueName: \"kubernetes.io/projected/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-kube-api-access-r56v2\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.821223 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-ceilometer-tls-certs\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.821304 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-log-httpd\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.821399 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-config-data\") pod \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\" (UID: \"f73614d7-c8d1-40bd-93dc-0bf5dcd82624\") " Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.822196 4675 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.823199 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.827840 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-kube-api-access-r56v2" (OuterVolumeSpecName: "kube-api-access-r56v2") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "kube-api-access-r56v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.830015 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-scripts" (OuterVolumeSpecName: "scripts") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.859732 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.888028 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.923213 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.923256 4675 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.923267 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r56v2\" (UniqueName: \"kubernetes.io/projected/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-kube-api-access-r56v2\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.923277 4675 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.923528 4675 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.923857 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:25 crc kubenswrapper[4675]: I0219 09:04:25.947782 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-config-data" (OuterVolumeSpecName: "config-data") pod "f73614d7-c8d1-40bd-93dc-0bf5dcd82624" (UID: "f73614d7-c8d1-40bd-93dc-0bf5dcd82624"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.024768 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.024803 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f73614d7-c8d1-40bd-93dc-0bf5dcd82624-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.089189 4675 generic.go:334] "Generic (PLEG): container finished" podID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerID="430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4" exitCode=0 Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.089234 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerDied","Data":"430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4"} Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.089271 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f73614d7-c8d1-40bd-93dc-0bf5dcd82624","Type":"ContainerDied","Data":"7b50301d5789a09faebca309110c13736879204e0a23f8e297086c6df2d0c726"} Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.089305 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.089307 4675 scope.go:117] "RemoveContainer" containerID="094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.103644 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.144897 4675 scope.go:117] "RemoveContainer" containerID="1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.153130 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.172964 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.197399 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.197900 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-notification-agent" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.197921 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-notification-agent" Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.197969 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-central-agent" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.197978 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-central-agent" Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.197998 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="sg-core" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.198005 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="sg-core" Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.198013 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="proxy-httpd" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.198019 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="proxy-httpd" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.198188 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-notification-agent" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.198198 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="proxy-httpd" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.198210 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="sg-core" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.198226 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" containerName="ceilometer-central-agent" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.200005 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.201925 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.202297 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.207757 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.232037 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.310945 4675 scope.go:117] "RemoveContainer" containerID="61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.338036 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhxdz\" (UniqueName: \"kubernetes.io/projected/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-kube-api-access-bhxdz\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.338397 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-scripts\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.338538 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-config-data\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.338644 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.338796 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-run-httpd\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.338894 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-log-httpd\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.338981 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.339300 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.346195 4675 scope.go:117] "RemoveContainer" containerID="430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.407110 4675 scope.go:117] "RemoveContainer" containerID="094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4" Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.410398 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4\": container with ID starting with 094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4 not found: ID does not exist" containerID="094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.410464 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4"} err="failed to get container status \"094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4\": rpc error: code = NotFound desc = could not find container \"094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4\": container with ID starting with 094d84fe52faff7b05f363af06c1927c1f6cf0fc0c0780035a61c794c72088d4 not found: ID does not exist" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.410487 4675 scope.go:117] "RemoveContainer" containerID="1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d" Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.410976 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d\": container with ID starting with 1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d not found: ID does not exist" containerID="1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.411043 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d"} err="failed to get container status \"1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d\": rpc error: code = NotFound desc = could not find container \"1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d\": container with ID starting with 1ec73b1f894e6be85077a8b88e398d29f6db55703a6145f34895c4f08646264d not found: ID does not exist" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.411079 4675 scope.go:117] "RemoveContainer" containerID="61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4" Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.411374 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4\": container with ID starting with 61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4 not found: ID does not exist" containerID="61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.411413 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4"} err="failed to get container status \"61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4\": rpc error: code = NotFound desc = could not find container \"61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4\": container with ID starting with 61230fe485b9e06058378b977c3b5d125783a7ebf9bea1869e99699dbeb7f4b4 not found: ID does not exist" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.411436 4675 scope.go:117] "RemoveContainer" containerID="430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4" Feb 19 09:04:26 crc kubenswrapper[4675]: E0219 09:04:26.411695 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4\": container with ID starting with 430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4 not found: ID does not exist" containerID="430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.411740 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4"} err="failed to get container status \"430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4\": rpc error: code = NotFound desc = could not find container \"430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4\": container with ID starting with 430fa31099286a886fe0ed015806369c03cc362df5356ffc70b1dce8620014b4 not found: ID does not exist" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.440473 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.440789 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-run-httpd\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.440875 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-log-httpd\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.440895 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.440924 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.440948 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhxdz\" (UniqueName: \"kubernetes.io/projected/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-kube-api-access-bhxdz\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.440967 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-scripts\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.441213 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-config-data\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.441941 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-log-httpd\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.441954 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-run-httpd\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.447211 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.448564 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-config-data\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.448901 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-scripts\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.450464 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.456283 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.461058 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhxdz\" (UniqueName: \"kubernetes.io/projected/f19360d2-5c69-4ad4-b4d5-9098dbf5bd84-kube-api-access-bhxdz\") pod \"ceilometer-0\" (UID: \"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84\") " pod="openstack/ceilometer-0" Feb 19 09:04:26 crc kubenswrapper[4675]: I0219 09:04:26.637464 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.092695 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.127615 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73abfd4a-24a9-4dc1-8064-1297d004af92" path="/var/lib/kubelet/pods/73abfd4a-24a9-4dc1-8064-1297d004af92/volumes" Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.128711 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f73614d7-c8d1-40bd-93dc-0bf5dcd82624" path="/var/lib/kubelet/pods/f73614d7-c8d1-40bd-93dc-0bf5dcd82624/volumes" Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.129588 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c28f0f-c32d-4114-b61e-592c108080e2","Type":"ContainerStarted","Data":"12ce3b6ee6fedb078c83521bcc283f4c21da203370b759d6167df98c5ce7dc90"} Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.129700 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c28f0f-c32d-4114-b61e-592c108080e2","Type":"ContainerStarted","Data":"6656e97646727c2ecc288166c2c5e5f43f708a6a30d77905f02ca6374ece0e4f"} Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.129725 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c28f0f-c32d-4114-b61e-592c108080e2","Type":"ContainerStarted","Data":"514c1b47d705c0d7cec7b1ac8283b53dfd2729b8778ed2925027bea259930e39"} Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.148132 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.148105698 podStartE2EDuration="2.148105698s" podCreationTimestamp="2026-02-19 09:04:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:04:27.133749183 +0000 UTC m=+1248.760839451" watchObservedRunningTime="2026-02-19 09:04:27.148105698 +0000 UTC m=+1248.775195966" Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.604855 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:27 crc kubenswrapper[4675]: I0219 09:04:27.627530 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.119552 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84","Type":"ContainerStarted","Data":"7452ffbc4df2e579163adf123041ac362e480f5d43253d60bc1662085a99351b"} Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.135671 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.334270 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-pxgzv"] Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.354503 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.356957 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pxgzv"] Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.358182 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.358549 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.486614 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-scripts\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.486703 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpfwh\" (UniqueName: \"kubernetes.io/projected/96629be8-e1e2-4cf3-8a9b-9df201341afb-kube-api-access-vpfwh\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.486863 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-config-data\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.486979 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.588517 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-config-data\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.588588 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.588692 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-scripts\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.588715 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpfwh\" (UniqueName: \"kubernetes.io/projected/96629be8-e1e2-4cf3-8a9b-9df201341afb-kube-api-access-vpfwh\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.596487 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-config-data\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.611421 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-scripts\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.611552 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.615101 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpfwh\" (UniqueName: \"kubernetes.io/projected/96629be8-e1e2-4cf3-8a9b-9df201341afb-kube-api-access-vpfwh\") pod \"nova-cell1-cell-mapping-pxgzv\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.660869 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.685738 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.762015 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vbjhx"] Feb 19 09:04:28 crc kubenswrapper[4675]: I0219 09:04:28.762305 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" podUID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerName="dnsmasq-dns" containerID="cri-o://1eb8570789ed39dac9f15de3c04ad5cdd9e3a27049c73a3af740861c59ac7368" gracePeriod=10 Feb 19 09:04:28 crc kubenswrapper[4675]: E0219 09:04:28.889128 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5a1bc99_3e8c_4078_a36f_853c6f785137.slice/crio-conmon-1eb8570789ed39dac9f15de3c04ad5cdd9e3a27049c73a3af740861c59ac7368.scope\": RecentStats: unable to find data in memory cache]" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.158448 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84","Type":"ContainerStarted","Data":"1d0816c6fc3951c28122c089d6aef17c74d1d75ece5507a41d851f73c9ad89f6"} Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.161007 4675 generic.go:334] "Generic (PLEG): container finished" podID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerID="1eb8570789ed39dac9f15de3c04ad5cdd9e3a27049c73a3af740861c59ac7368" exitCode=0 Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.163277 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" event={"ID":"f5a1bc99-3e8c-4078-a36f-853c6f785137","Type":"ContainerDied","Data":"1eb8570789ed39dac9f15de3c04ad5cdd9e3a27049c73a3af740861c59ac7368"} Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.264391 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pxgzv"] Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.370829 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.513124 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-nb\") pod \"f5a1bc99-3e8c-4078-a36f-853c6f785137\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.513187 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-sb\") pod \"f5a1bc99-3e8c-4078-a36f-853c6f785137\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.513216 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-svc\") pod \"f5a1bc99-3e8c-4078-a36f-853c6f785137\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.513357 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-config\") pod \"f5a1bc99-3e8c-4078-a36f-853c6f785137\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.513413 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5zxc\" (UniqueName: \"kubernetes.io/projected/f5a1bc99-3e8c-4078-a36f-853c6f785137-kube-api-access-f5zxc\") pod \"f5a1bc99-3e8c-4078-a36f-853c6f785137\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.513517 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-swift-storage-0\") pod \"f5a1bc99-3e8c-4078-a36f-853c6f785137\" (UID: \"f5a1bc99-3e8c-4078-a36f-853c6f785137\") " Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.518094 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a1bc99-3e8c-4078-a36f-853c6f785137-kube-api-access-f5zxc" (OuterVolumeSpecName: "kube-api-access-f5zxc") pod "f5a1bc99-3e8c-4078-a36f-853c6f785137" (UID: "f5a1bc99-3e8c-4078-a36f-853c6f785137"). InnerVolumeSpecName "kube-api-access-f5zxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.565007 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f5a1bc99-3e8c-4078-a36f-853c6f785137" (UID: "f5a1bc99-3e8c-4078-a36f-853c6f785137"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.565034 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f5a1bc99-3e8c-4078-a36f-853c6f785137" (UID: "f5a1bc99-3e8c-4078-a36f-853c6f785137"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.566023 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-config" (OuterVolumeSpecName: "config") pod "f5a1bc99-3e8c-4078-a36f-853c6f785137" (UID: "f5a1bc99-3e8c-4078-a36f-853c6f785137"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.569365 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f5a1bc99-3e8c-4078-a36f-853c6f785137" (UID: "f5a1bc99-3e8c-4078-a36f-853c6f785137"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.569569 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f5a1bc99-3e8c-4078-a36f-853c6f785137" (UID: "f5a1bc99-3e8c-4078-a36f-853c6f785137"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.616405 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.617330 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.617510 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.617677 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.617824 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5zxc\" (UniqueName: \"kubernetes.io/projected/f5a1bc99-3e8c-4078-a36f-853c6f785137-kube-api-access-f5zxc\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:29 crc kubenswrapper[4675]: I0219 09:04:29.617979 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5a1bc99-3e8c-4078-a36f-853c6f785137-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.174972 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84","Type":"ContainerStarted","Data":"e6779ade6ae85109ea342c3aee563f30979ecf3eeef2290ae3b7a999701ae5d5"} Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.175020 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84","Type":"ContainerStarted","Data":"988f0669e80954062fcb08f6d4455dabb49877aa591ecc8a514bb2eb16788796"} Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.178989 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.179003 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-vbjhx" event={"ID":"f5a1bc99-3e8c-4078-a36f-853c6f785137","Type":"ContainerDied","Data":"bfb83dbdd46f546da01009b0f26ce051e727d6536965774b0ff8750ccbc89441"} Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.179239 4675 scope.go:117] "RemoveContainer" containerID="1eb8570789ed39dac9f15de3c04ad5cdd9e3a27049c73a3af740861c59ac7368" Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.180977 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pxgzv" event={"ID":"96629be8-e1e2-4cf3-8a9b-9df201341afb","Type":"ContainerStarted","Data":"f5d7e5b8484626896ef08e590e60ad383e0ac383870f47f9316aca3e5a31e62e"} Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.181019 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pxgzv" event={"ID":"96629be8-e1e2-4cf3-8a9b-9df201341afb","Type":"ContainerStarted","Data":"cf8441825423bf2e9366e57d2db5c3ef21b66e5c774f22bd33c908408801e549"} Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.198751 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-pxgzv" podStartSLOduration=2.198730985 podStartE2EDuration="2.198730985s" podCreationTimestamp="2026-02-19 09:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:04:30.19814992 +0000 UTC m=+1251.825240198" watchObservedRunningTime="2026-02-19 09:04:30.198730985 +0000 UTC m=+1251.825821253" Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.203262 4675 scope.go:117] "RemoveContainer" containerID="adbe0d9e78fcc385581c1b309f49a120d8f2e65e17554a1897145951dece975f" Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.234030 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vbjhx"] Feb 19 09:04:30 crc kubenswrapper[4675]: I0219 09:04:30.241240 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-vbjhx"] Feb 19 09:04:31 crc kubenswrapper[4675]: I0219 09:04:31.116989 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a1bc99-3e8c-4078-a36f-853c6f785137" path="/var/lib/kubelet/pods/f5a1bc99-3e8c-4078-a36f-853c6f785137/volumes" Feb 19 09:04:32 crc kubenswrapper[4675]: I0219 09:04:32.219278 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f19360d2-5c69-4ad4-b4d5-9098dbf5bd84","Type":"ContainerStarted","Data":"13399030bfc7fc437b2a8c607b30d2414ef0cf6e9c833d47d72e873fe051b146"} Feb 19 09:04:32 crc kubenswrapper[4675]: I0219 09:04:32.220031 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 19 09:04:32 crc kubenswrapper[4675]: I0219 09:04:32.250965 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.090688067 podStartE2EDuration="6.250945738s" podCreationTimestamp="2026-02-19 09:04:26 +0000 UTC" firstStartedPulling="2026-02-19 09:04:27.097320134 +0000 UTC m=+1248.724410402" lastFinishedPulling="2026-02-19 09:04:31.257577805 +0000 UTC m=+1252.884668073" observedRunningTime="2026-02-19 09:04:32.243548285 +0000 UTC m=+1253.870638873" watchObservedRunningTime="2026-02-19 09:04:32.250945738 +0000 UTC m=+1253.878035996" Feb 19 09:04:35 crc kubenswrapper[4675]: I0219 09:04:35.254051 4675 generic.go:334] "Generic (PLEG): container finished" podID="96629be8-e1e2-4cf3-8a9b-9df201341afb" containerID="f5d7e5b8484626896ef08e590e60ad383e0ac383870f47f9316aca3e5a31e62e" exitCode=0 Feb 19 09:04:35 crc kubenswrapper[4675]: I0219 09:04:35.254546 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pxgzv" event={"ID":"96629be8-e1e2-4cf3-8a9b-9df201341afb","Type":"ContainerDied","Data":"f5d7e5b8484626896ef08e590e60ad383e0ac383870f47f9316aca3e5a31e62e"} Feb 19 09:04:35 crc kubenswrapper[4675]: I0219 09:04:35.557013 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:04:35 crc kubenswrapper[4675]: I0219 09:04:35.557112 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.572810 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.572868 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.708875 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.831197 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-config-data\") pod \"96629be8-e1e2-4cf3-8a9b-9df201341afb\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.831557 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-combined-ca-bundle\") pod \"96629be8-e1e2-4cf3-8a9b-9df201341afb\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.831585 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-scripts\") pod \"96629be8-e1e2-4cf3-8a9b-9df201341afb\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.831824 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpfwh\" (UniqueName: \"kubernetes.io/projected/96629be8-e1e2-4cf3-8a9b-9df201341afb-kube-api-access-vpfwh\") pod \"96629be8-e1e2-4cf3-8a9b-9df201341afb\" (UID: \"96629be8-e1e2-4cf3-8a9b-9df201341afb\") " Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.847015 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-scripts" (OuterVolumeSpecName: "scripts") pod "96629be8-e1e2-4cf3-8a9b-9df201341afb" (UID: "96629be8-e1e2-4cf3-8a9b-9df201341afb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.848762 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96629be8-e1e2-4cf3-8a9b-9df201341afb-kube-api-access-vpfwh" (OuterVolumeSpecName: "kube-api-access-vpfwh") pod "96629be8-e1e2-4cf3-8a9b-9df201341afb" (UID: "96629be8-e1e2-4cf3-8a9b-9df201341afb"). InnerVolumeSpecName "kube-api-access-vpfwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.867908 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96629be8-e1e2-4cf3-8a9b-9df201341afb" (UID: "96629be8-e1e2-4cf3-8a9b-9df201341afb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.877431 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-config-data" (OuterVolumeSpecName: "config-data") pod "96629be8-e1e2-4cf3-8a9b-9df201341afb" (UID: "96629be8-e1e2-4cf3-8a9b-9df201341afb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.934285 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpfwh\" (UniqueName: \"kubernetes.io/projected/96629be8-e1e2-4cf3-8a9b-9df201341afb-kube-api-access-vpfwh\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.934654 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.934668 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:36 crc kubenswrapper[4675]: I0219 09:04:36.934679 4675 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96629be8-e1e2-4cf3-8a9b-9df201341afb-scripts\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.276381 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pxgzv" event={"ID":"96629be8-e1e2-4cf3-8a9b-9df201341afb","Type":"ContainerDied","Data":"cf8441825423bf2e9366e57d2db5c3ef21b66e5c774f22bd33c908408801e549"} Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.276427 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf8441825423bf2e9366e57d2db5c3ef21b66e5c774f22bd33c908408801e549" Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.276451 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pxgzv" Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.466472 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.466801 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6a3bb7b0-46ad-488d-b4c1-7e486d09866f" containerName="nova-scheduler-scheduler" containerID="cri-o://e2afacffb7ba83062ad36f29920716477e238540a693024edf571a4430ad8c05" gracePeriod=30 Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.484436 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.484778 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-log" containerID="cri-o://6656e97646727c2ecc288166c2c5e5f43f708a6a30d77905f02ca6374ece0e4f" gracePeriod=30 Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.485233 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-api" containerID="cri-o://12ce3b6ee6fedb078c83521bcc283f4c21da203370b759d6167df98c5ce7dc90" gracePeriod=30 Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.498696 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.498969 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-log" containerID="cri-o://cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4" gracePeriod=30 Feb 19 09:04:37 crc kubenswrapper[4675]: I0219 09:04:37.499438 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-metadata" containerID="cri-o://f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f" gracePeriod=30 Feb 19 09:04:38 crc kubenswrapper[4675]: I0219 09:04:38.287092 4675 generic.go:334] "Generic (PLEG): container finished" podID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerID="cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4" exitCode=143 Feb 19 09:04:38 crc kubenswrapper[4675]: I0219 09:04:38.287176 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"322bc0cb-2559-4c88-93a1-01a3cc10c29c","Type":"ContainerDied","Data":"cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4"} Feb 19 09:04:38 crc kubenswrapper[4675]: I0219 09:04:38.288816 4675 generic.go:334] "Generic (PLEG): container finished" podID="78c28f0f-c32d-4114-b61e-592c108080e2" containerID="6656e97646727c2ecc288166c2c5e5f43f708a6a30d77905f02ca6374ece0e4f" exitCode=143 Feb 19 09:04:38 crc kubenswrapper[4675]: I0219 09:04:38.288841 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c28f0f-c32d-4114-b61e-592c108080e2","Type":"ContainerDied","Data":"6656e97646727c2ecc288166c2c5e5f43f708a6a30d77905f02ca6374ece0e4f"} Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.305569 4675 generic.go:334] "Generic (PLEG): container finished" podID="6a3bb7b0-46ad-488d-b4c1-7e486d09866f" containerID="e2afacffb7ba83062ad36f29920716477e238540a693024edf571a4430ad8c05" exitCode=0 Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.305659 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a3bb7b0-46ad-488d-b4c1-7e486d09866f","Type":"ContainerDied","Data":"e2afacffb7ba83062ad36f29920716477e238540a693024edf571a4430ad8c05"} Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.575936 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.688552 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-combined-ca-bundle\") pod \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.688744 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-config-data\") pod \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.688827 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2frw8\" (UniqueName: \"kubernetes.io/projected/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-kube-api-access-2frw8\") pod \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\" (UID: \"6a3bb7b0-46ad-488d-b4c1-7e486d09866f\") " Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.700753 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-kube-api-access-2frw8" (OuterVolumeSpecName: "kube-api-access-2frw8") pod "6a3bb7b0-46ad-488d-b4c1-7e486d09866f" (UID: "6a3bb7b0-46ad-488d-b4c1-7e486d09866f"). InnerVolumeSpecName "kube-api-access-2frw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.718034 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-config-data" (OuterVolumeSpecName: "config-data") pod "6a3bb7b0-46ad-488d-b4c1-7e486d09866f" (UID: "6a3bb7b0-46ad-488d-b4c1-7e486d09866f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.719527 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a3bb7b0-46ad-488d-b4c1-7e486d09866f" (UID: "6a3bb7b0-46ad-488d-b4c1-7e486d09866f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.797415 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.797451 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:39 crc kubenswrapper[4675]: I0219 09:04:39.797465 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2frw8\" (UniqueName: \"kubernetes.io/projected/6a3bb7b0-46ad-488d-b4c1-7e486d09866f-kube-api-access-2frw8\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.317851 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a3bb7b0-46ad-488d-b4c1-7e486d09866f","Type":"ContainerDied","Data":"0dcb7fd3db27d8ab37ffd08cee841351c6bbe7d05a7b5e73bf7ebd31d3d72258"} Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.317914 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.317918 4675 scope.go:117] "RemoveContainer" containerID="e2afacffb7ba83062ad36f29920716477e238540a693024edf571a4430ad8c05" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.351798 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.360202 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.381449 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:04:40 crc kubenswrapper[4675]: E0219 09:04:40.382245 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerName="init" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.382269 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerName="init" Feb 19 09:04:40 crc kubenswrapper[4675]: E0219 09:04:40.382303 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96629be8-e1e2-4cf3-8a9b-9df201341afb" containerName="nova-manage" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.382311 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="96629be8-e1e2-4cf3-8a9b-9df201341afb" containerName="nova-manage" Feb 19 09:04:40 crc kubenswrapper[4675]: E0219 09:04:40.382326 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerName="dnsmasq-dns" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.382334 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerName="dnsmasq-dns" Feb 19 09:04:40 crc kubenswrapper[4675]: E0219 09:04:40.382360 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3bb7b0-46ad-488d-b4c1-7e486d09866f" containerName="nova-scheduler-scheduler" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.382368 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3bb7b0-46ad-488d-b4c1-7e486d09866f" containerName="nova-scheduler-scheduler" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.382605 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3bb7b0-46ad-488d-b4c1-7e486d09866f" containerName="nova-scheduler-scheduler" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.382663 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a1bc99-3e8c-4078-a36f-853c6f785137" containerName="dnsmasq-dns" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.382681 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="96629be8-e1e2-4cf3-8a9b-9df201341afb" containerName="nova-manage" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.383520 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.386518 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.405906 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.411805 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n849\" (UniqueName: \"kubernetes.io/projected/de054e7d-f903-4b96-8394-7d12f6cfdb98-kube-api-access-7n849\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.411855 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de054e7d-f903-4b96-8394-7d12f6cfdb98-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.411956 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de054e7d-f903-4b96-8394-7d12f6cfdb98-config-data\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.514031 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n849\" (UniqueName: \"kubernetes.io/projected/de054e7d-f903-4b96-8394-7d12f6cfdb98-kube-api-access-7n849\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.514512 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de054e7d-f903-4b96-8394-7d12f6cfdb98-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.514667 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de054e7d-f903-4b96-8394-7d12f6cfdb98-config-data\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.520567 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de054e7d-f903-4b96-8394-7d12f6cfdb98-config-data\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.520961 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de054e7d-f903-4b96-8394-7d12f6cfdb98-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.530869 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n849\" (UniqueName: \"kubernetes.io/projected/de054e7d-f903-4b96-8394-7d12f6cfdb98-kube-api-access-7n849\") pod \"nova-scheduler-0\" (UID: \"de054e7d-f903-4b96-8394-7d12f6cfdb98\") " pod="openstack/nova-scheduler-0" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.628229 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:34742->10.217.0.199:8775: read: connection reset by peer" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.628229 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:34736->10.217.0.199:8775: read: connection reset by peer" Feb 19 09:04:40 crc kubenswrapper[4675]: I0219 09:04:40.724716 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.110056 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.116449 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3bb7b0-46ad-488d-b4c1-7e486d09866f" path="/var/lib/kubelet/pods/6a3bb7b0-46ad-488d-b4c1-7e486d09866f/volumes" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.237146 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 19 09:04:41 crc kubenswrapper[4675]: W0219 09:04:41.245762 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde054e7d_f903_4b96_8394_7d12f6cfdb98.slice/crio-6cc449438ccf5c5fbdcd37bf61c1d94d7e68aed40ad907dccc5be9eea8fae2e6 WatchSource:0}: Error finding container 6cc449438ccf5c5fbdcd37bf61c1d94d7e68aed40ad907dccc5be9eea8fae2e6: Status 404 returned error can't find the container with id 6cc449438ccf5c5fbdcd37bf61c1d94d7e68aed40ad907dccc5be9eea8fae2e6 Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.252081 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwr6g\" (UniqueName: \"kubernetes.io/projected/322bc0cb-2559-4c88-93a1-01a3cc10c29c-kube-api-access-cwr6g\") pod \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.252139 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322bc0cb-2559-4c88-93a1-01a3cc10c29c-logs\") pod \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.252185 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-config-data\") pod \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.252215 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-combined-ca-bundle\") pod \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.252422 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-nova-metadata-tls-certs\") pod \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\" (UID: \"322bc0cb-2559-4c88-93a1-01a3cc10c29c\") " Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.253016 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322bc0cb-2559-4c88-93a1-01a3cc10c29c-logs" (OuterVolumeSpecName: "logs") pod "322bc0cb-2559-4c88-93a1-01a3cc10c29c" (UID: "322bc0cb-2559-4c88-93a1-01a3cc10c29c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.256809 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/322bc0cb-2559-4c88-93a1-01a3cc10c29c-kube-api-access-cwr6g" (OuterVolumeSpecName: "kube-api-access-cwr6g") pod "322bc0cb-2559-4c88-93a1-01a3cc10c29c" (UID: "322bc0cb-2559-4c88-93a1-01a3cc10c29c"). InnerVolumeSpecName "kube-api-access-cwr6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.284671 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-config-data" (OuterVolumeSpecName: "config-data") pod "322bc0cb-2559-4c88-93a1-01a3cc10c29c" (UID: "322bc0cb-2559-4c88-93a1-01a3cc10c29c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.292000 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "322bc0cb-2559-4c88-93a1-01a3cc10c29c" (UID: "322bc0cb-2559-4c88-93a1-01a3cc10c29c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.328426 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "322bc0cb-2559-4c88-93a1-01a3cc10c29c" (UID: "322bc0cb-2559-4c88-93a1-01a3cc10c29c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.339853 4675 generic.go:334] "Generic (PLEG): container finished" podID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerID="f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f" exitCode=0 Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.340020 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.341006 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"322bc0cb-2559-4c88-93a1-01a3cc10c29c","Type":"ContainerDied","Data":"f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f"} Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.341052 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"322bc0cb-2559-4c88-93a1-01a3cc10c29c","Type":"ContainerDied","Data":"83aa733e217322b62971bab6c293bd91522afec07b9aea215bd976f4f5b3d325"} Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.341071 4675 scope.go:117] "RemoveContainer" containerID="f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.342860 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"de054e7d-f903-4b96-8394-7d12f6cfdb98","Type":"ContainerStarted","Data":"6cc449438ccf5c5fbdcd37bf61c1d94d7e68aed40ad907dccc5be9eea8fae2e6"} Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.354312 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwr6g\" (UniqueName: \"kubernetes.io/projected/322bc0cb-2559-4c88-93a1-01a3cc10c29c-kube-api-access-cwr6g\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.354344 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322bc0cb-2559-4c88-93a1-01a3cc10c29c-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.354354 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.354363 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.354373 4675 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/322bc0cb-2559-4c88-93a1-01a3cc10c29c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.366207 4675 scope.go:117] "RemoveContainer" containerID="cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.393078 4675 scope.go:117] "RemoveContainer" containerID="f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f" Feb 19 09:04:41 crc kubenswrapper[4675]: E0219 09:04:41.395055 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f\": container with ID starting with f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f not found: ID does not exist" containerID="f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.395099 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f"} err="failed to get container status \"f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f\": rpc error: code = NotFound desc = could not find container \"f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f\": container with ID starting with f9540902f726f3c918d5b0c01a57e0cc0e632ff5fbcf9f12a89b14e4f97aea4f not found: ID does not exist" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.395125 4675 scope.go:117] "RemoveContainer" containerID="cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4" Feb 19 09:04:41 crc kubenswrapper[4675]: E0219 09:04:41.395418 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4\": container with ID starting with cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4 not found: ID does not exist" containerID="cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.395474 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4"} err="failed to get container status \"cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4\": rpc error: code = NotFound desc = could not find container \"cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4\": container with ID starting with cd7f699fc5dc5228c57da1111908e8d7cd8855b4b8a517f165359bfca572c0c4 not found: ID does not exist" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.400619 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.411894 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.430417 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:04:41 crc kubenswrapper[4675]: E0219 09:04:41.430902 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-metadata" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.430924 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-metadata" Feb 19 09:04:41 crc kubenswrapper[4675]: E0219 09:04:41.430944 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-log" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.430953 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-log" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.431169 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-metadata" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.431196 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" containerName="nova-metadata-log" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.432176 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.435353 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.435793 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.456175 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-config-data\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.456518 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.456671 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.456807 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78796291-e993-4cea-90df-f44081e71bf9-logs\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.456940 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4knt2\" (UniqueName: \"kubernetes.io/projected/78796291-e993-4cea-90df-f44081e71bf9-kube-api-access-4knt2\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.462741 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.558115 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78796291-e993-4cea-90df-f44081e71bf9-logs\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.558207 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4knt2\" (UniqueName: \"kubernetes.io/projected/78796291-e993-4cea-90df-f44081e71bf9-kube-api-access-4knt2\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.558325 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-config-data\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.558365 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.558418 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.559053 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78796291-e993-4cea-90df-f44081e71bf9-logs\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.562862 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.563665 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-config-data\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.565705 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78796291-e993-4cea-90df-f44081e71bf9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.576194 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4knt2\" (UniqueName: \"kubernetes.io/projected/78796291-e993-4cea-90df-f44081e71bf9-kube-api-access-4knt2\") pod \"nova-metadata-0\" (UID: \"78796291-e993-4cea-90df-f44081e71bf9\") " pod="openstack/nova-metadata-0" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.743411 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.743472 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.743516 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.744305 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29c6cf62bac0dde894372357f7b07e5a365b8cc260e995aa58cda63d08dcff6f"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.744370 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://29c6cf62bac0dde894372357f7b07e5a365b8cc260e995aa58cda63d08dcff6f" gracePeriod=600 Feb 19 09:04:41 crc kubenswrapper[4675]: I0219 09:04:41.758520 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.233774 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 19 09:04:42 crc kubenswrapper[4675]: W0219 09:04:42.251361 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78796291_e993_4cea_90df_f44081e71bf9.slice/crio-0d73ce5f60d79075856510a99a6d1cee6ce0021c171a14bcdf4f1b1a85fb8886 WatchSource:0}: Error finding container 0d73ce5f60d79075856510a99a6d1cee6ce0021c171a14bcdf4f1b1a85fb8886: Status 404 returned error can't find the container with id 0d73ce5f60d79075856510a99a6d1cee6ce0021c171a14bcdf4f1b1a85fb8886 Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.379822 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"de054e7d-f903-4b96-8394-7d12f6cfdb98","Type":"ContainerStarted","Data":"1cc6ab9c2e23bbce00dec434023900bfec6f93f2692bed6bd39b9ed5e1e8d22d"} Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.385302 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="29c6cf62bac0dde894372357f7b07e5a365b8cc260e995aa58cda63d08dcff6f" exitCode=0 Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.385373 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"29c6cf62bac0dde894372357f7b07e5a365b8cc260e995aa58cda63d08dcff6f"} Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.385398 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"d8cec9b5ad0c194b9e5ae1c687b8aae050e8cf77a0f404f058bb11195c4e24e3"} Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.385416 4675 scope.go:117] "RemoveContainer" containerID="ca5833a715b96b50d8038acbecde68c582755da122b5dea38657e570ff140d31" Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.391540 4675 generic.go:334] "Generic (PLEG): container finished" podID="78c28f0f-c32d-4114-b61e-592c108080e2" containerID="12ce3b6ee6fedb078c83521bcc283f4c21da203370b759d6167df98c5ce7dc90" exitCode=0 Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.391656 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c28f0f-c32d-4114-b61e-592c108080e2","Type":"ContainerDied","Data":"12ce3b6ee6fedb078c83521bcc283f4c21da203370b759d6167df98c5ce7dc90"} Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.393380 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.397124 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78796291-e993-4cea-90df-f44081e71bf9","Type":"ContainerStarted","Data":"0d73ce5f60d79075856510a99a6d1cee6ce0021c171a14bcdf4f1b1a85fb8886"} Feb 19 09:04:42 crc kubenswrapper[4675]: I0219 09:04:42.400839 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.400764665 podStartE2EDuration="2.400764665s" podCreationTimestamp="2026-02-19 09:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:04:42.397970442 +0000 UTC m=+1264.025060710" watchObservedRunningTime="2026-02-19 09:04:42.400764665 +0000 UTC m=+1264.027854933" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.580331 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c28f0f-c32d-4114-b61e-592c108080e2-logs\") pod \"78c28f0f-c32d-4114-b61e-592c108080e2\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.580780 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xhhj\" (UniqueName: \"kubernetes.io/projected/78c28f0f-c32d-4114-b61e-592c108080e2-kube-api-access-9xhhj\") pod \"78c28f0f-c32d-4114-b61e-592c108080e2\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.580805 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-internal-tls-certs\") pod \"78c28f0f-c32d-4114-b61e-592c108080e2\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.580926 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-combined-ca-bundle\") pod \"78c28f0f-c32d-4114-b61e-592c108080e2\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.580984 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-public-tls-certs\") pod \"78c28f0f-c32d-4114-b61e-592c108080e2\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.581101 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-config-data\") pod \"78c28f0f-c32d-4114-b61e-592c108080e2\" (UID: \"78c28f0f-c32d-4114-b61e-592c108080e2\") " Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.581097 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78c28f0f-c32d-4114-b61e-592c108080e2-logs" (OuterVolumeSpecName: "logs") pod "78c28f0f-c32d-4114-b61e-592c108080e2" (UID: "78c28f0f-c32d-4114-b61e-592c108080e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.602765 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c28f0f-c32d-4114-b61e-592c108080e2-kube-api-access-9xhhj" (OuterVolumeSpecName: "kube-api-access-9xhhj") pod "78c28f0f-c32d-4114-b61e-592c108080e2" (UID: "78c28f0f-c32d-4114-b61e-592c108080e2"). InnerVolumeSpecName "kube-api-access-9xhhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.639701 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-config-data" (OuterVolumeSpecName: "config-data") pod "78c28f0f-c32d-4114-b61e-592c108080e2" (UID: "78c28f0f-c32d-4114-b61e-592c108080e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.640614 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78c28f0f-c32d-4114-b61e-592c108080e2" (UID: "78c28f0f-c32d-4114-b61e-592c108080e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.663766 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "78c28f0f-c32d-4114-b61e-592c108080e2" (UID: "78c28f0f-c32d-4114-b61e-592c108080e2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.665539 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "78c28f0f-c32d-4114-b61e-592c108080e2" (UID: "78c28f0f-c32d-4114-b61e-592c108080e2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.683142 4675 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.683173 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.683182 4675 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c28f0f-c32d-4114-b61e-592c108080e2-logs\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.683194 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xhhj\" (UniqueName: \"kubernetes.io/projected/78c28f0f-c32d-4114-b61e-592c108080e2-kube-api-access-9xhhj\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.683205 4675 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:42.683214 4675 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c28f0f-c32d-4114-b61e-592c108080e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.114894 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="322bc0cb-2559-4c88-93a1-01a3cc10c29c" path="/var/lib/kubelet/pods/322bc0cb-2559-4c88-93a1-01a3cc10c29c/volumes" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.409793 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c28f0f-c32d-4114-b61e-592c108080e2","Type":"ContainerDied","Data":"514c1b47d705c0d7cec7b1ac8283b53dfd2729b8778ed2925027bea259930e39"} Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.409837 4675 scope.go:117] "RemoveContainer" containerID="12ce3b6ee6fedb078c83521bcc283f4c21da203370b759d6167df98c5ce7dc90" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.409931 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.415490 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78796291-e993-4cea-90df-f44081e71bf9","Type":"ContainerStarted","Data":"05ca42a094c0d661690d273a402913a157f641afd36cc748a71cb5ca27b54c5c"} Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.415521 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78796291-e993-4cea-90df-f44081e71bf9","Type":"ContainerStarted","Data":"7527a926e190f0ef910aa5f9ddb2fe8339bebf97190d663defd2b282c2033911"} Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.436402 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.445744 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.459107 4675 scope.go:117] "RemoveContainer" containerID="6656e97646727c2ecc288166c2c5e5f43f708a6a30d77905f02ca6374ece0e4f" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.464147 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:43 crc kubenswrapper[4675]: E0219 09:04:43.464586 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-log" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.464602 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-log" Feb 19 09:04:43 crc kubenswrapper[4675]: E0219 09:04:43.464722 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-api" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.464731 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-api" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.464972 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-log" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.464987 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" containerName="nova-api-api" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.466127 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4661161529999998 podStartE2EDuration="2.466116153s" podCreationTimestamp="2026-02-19 09:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:04:43.456631606 +0000 UTC m=+1265.083721874" watchObservedRunningTime="2026-02-19 09:04:43.466116153 +0000 UTC m=+1265.093206421" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.466270 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.471310 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.471515 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.473876 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.506852 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.600676 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-public-tls-certs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.600717 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-config-data\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.600756 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.600785 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.600870 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5lzh\" (UniqueName: \"kubernetes.io/projected/657eb3f7-7b32-44db-b26b-c070b619cee7-kube-api-access-s5lzh\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.600927 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/657eb3f7-7b32-44db-b26b-c070b619cee7-logs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.703234 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/657eb3f7-7b32-44db-b26b-c070b619cee7-logs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.703331 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-public-tls-certs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.703365 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-config-data\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.703400 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.703438 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.703527 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5lzh\" (UniqueName: \"kubernetes.io/projected/657eb3f7-7b32-44db-b26b-c070b619cee7-kube-api-access-s5lzh\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.703741 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/657eb3f7-7b32-44db-b26b-c070b619cee7-logs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.709017 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-public-tls-certs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.709387 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.709409 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-config-data\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.709873 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/657eb3f7-7b32-44db-b26b-c070b619cee7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.720722 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5lzh\" (UniqueName: \"kubernetes.io/projected/657eb3f7-7b32-44db-b26b-c070b619cee7-kube-api-access-s5lzh\") pod \"nova-api-0\" (UID: \"657eb3f7-7b32-44db-b26b-c070b619cee7\") " pod="openstack/nova-api-0" Feb 19 09:04:43 crc kubenswrapper[4675]: I0219 09:04:43.798581 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 19 09:04:44 crc kubenswrapper[4675]: I0219 09:04:44.249012 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 19 09:04:44 crc kubenswrapper[4675]: W0219 09:04:44.253746 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod657eb3f7_7b32_44db_b26b_c070b619cee7.slice/crio-bc99cff96615d42b5676045fe28761fc70c9f2c2c34755ac59d5ac4b598e1683 WatchSource:0}: Error finding container bc99cff96615d42b5676045fe28761fc70c9f2c2c34755ac59d5ac4b598e1683: Status 404 returned error can't find the container with id bc99cff96615d42b5676045fe28761fc70c9f2c2c34755ac59d5ac4b598e1683 Feb 19 09:04:44 crc kubenswrapper[4675]: I0219 09:04:44.428349 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"657eb3f7-7b32-44db-b26b-c070b619cee7","Type":"ContainerStarted","Data":"b4e46143123ff6ce4c5b108e3c12cb6a7186ace4e9f84d4b44cf619d813ab1e5"} Feb 19 09:04:44 crc kubenswrapper[4675]: I0219 09:04:44.428956 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"657eb3f7-7b32-44db-b26b-c070b619cee7","Type":"ContainerStarted","Data":"bc99cff96615d42b5676045fe28761fc70c9f2c2c34755ac59d5ac4b598e1683"} Feb 19 09:04:45 crc kubenswrapper[4675]: I0219 09:04:45.116008 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c28f0f-c32d-4114-b61e-592c108080e2" path="/var/lib/kubelet/pods/78c28f0f-c32d-4114-b61e-592c108080e2/volumes" Feb 19 09:04:45 crc kubenswrapper[4675]: I0219 09:04:45.472861 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"657eb3f7-7b32-44db-b26b-c070b619cee7","Type":"ContainerStarted","Data":"6f57953293c3678b359c1bb8f9613a9621c801381839ab8a3fa56d678f630a86"} Feb 19 09:04:45 crc kubenswrapper[4675]: I0219 09:04:45.494982 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.494963828 podStartE2EDuration="2.494963828s" podCreationTimestamp="2026-02-19 09:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:04:45.489929137 +0000 UTC m=+1267.117019415" watchObservedRunningTime="2026-02-19 09:04:45.494963828 +0000 UTC m=+1267.122054096" Feb 19 09:04:45 crc kubenswrapper[4675]: I0219 09:04:45.725580 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 19 09:04:46 crc kubenswrapper[4675]: I0219 09:04:46.759828 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 19 09:04:46 crc kubenswrapper[4675]: I0219 09:04:46.760344 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 19 09:04:50 crc kubenswrapper[4675]: I0219 09:04:50.725281 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 19 09:04:50 crc kubenswrapper[4675]: I0219 09:04:50.758528 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 19 09:04:51 crc kubenswrapper[4675]: I0219 09:04:51.555520 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 19 09:04:51 crc kubenswrapper[4675]: I0219 09:04:51.759707 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 19 09:04:51 crc kubenswrapper[4675]: I0219 09:04:51.760144 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 19 09:04:52 crc kubenswrapper[4675]: I0219 09:04:52.777806 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="78796291-e993-4cea-90df-f44081e71bf9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:52 crc kubenswrapper[4675]: I0219 09:04:52.777806 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="78796291-e993-4cea-90df-f44081e71bf9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:53 crc kubenswrapper[4675]: I0219 09:04:53.798865 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:04:53 crc kubenswrapper[4675]: I0219 09:04:53.799225 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 19 09:04:54 crc kubenswrapper[4675]: I0219 09:04:54.810823 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="657eb3f7-7b32-44db-b26b-c070b619cee7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:54 crc kubenswrapper[4675]: I0219 09:04:54.811919 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="657eb3f7-7b32-44db-b26b-c070b619cee7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 19 09:04:56 crc kubenswrapper[4675]: I0219 09:04:56.647148 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 19 09:05:01 crc kubenswrapper[4675]: I0219 09:05:01.775571 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 19 09:05:01 crc kubenswrapper[4675]: I0219 09:05:01.783851 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 19 09:05:01 crc kubenswrapper[4675]: I0219 09:05:01.785856 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 19 09:05:02 crc kubenswrapper[4675]: I0219 09:05:02.638077 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 19 09:05:03 crc kubenswrapper[4675]: I0219 09:05:03.808966 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 19 09:05:03 crc kubenswrapper[4675]: I0219 09:05:03.809937 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 19 09:05:03 crc kubenswrapper[4675]: I0219 09:05:03.812353 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 19 09:05:03 crc kubenswrapper[4675]: I0219 09:05:03.815910 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 19 09:05:04 crc kubenswrapper[4675]: I0219 09:05:04.652456 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 19 09:05:04 crc kubenswrapper[4675]: I0219 09:05:04.661012 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 19 09:05:12 crc kubenswrapper[4675]: I0219 09:05:12.555475 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 09:05:14 crc kubenswrapper[4675]: I0219 09:05:14.117197 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 09:05:16 crc kubenswrapper[4675]: I0219 09:05:16.996221 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerName="rabbitmq" containerID="cri-o://f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e" gracePeriod=604796 Feb 19 09:05:18 crc kubenswrapper[4675]: I0219 09:05:18.355324 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerName="rabbitmq" containerID="cri-o://db61585170f749410d45a8cfd2979a86557f0fc24ab094b737b4fa31e9bf3613" gracePeriod=604796 Feb 19 09:05:22 crc kubenswrapper[4675]: I0219 09:05:22.797521 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.106061 4675 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.604242 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.747572 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-erlang-cookie\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.747654 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-config-data\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.747859 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93dbb123-819d-44f7-80f5-8c38c639f5f4-pod-info\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.747902 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-confd\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.747934 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.747957 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93dbb123-819d-44f7-80f5-8c38c639f5f4-erlang-cookie-secret\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.748000 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-server-conf\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.748037 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-plugins-conf\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.748147 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-plugins\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.748171 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-tls\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.748216 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bxqz\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-kube-api-access-6bxqz\") pod \"93dbb123-819d-44f7-80f5-8c38c639f5f4\" (UID: \"93dbb123-819d-44f7-80f5-8c38c639f5f4\") " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.755600 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.756762 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.757111 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/93dbb123-819d-44f7-80f5-8c38c639f5f4-pod-info" (OuterVolumeSpecName: "pod-info") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.757052 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.758506 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.760221 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-kube-api-access-6bxqz" (OuterVolumeSpecName: "kube-api-access-6bxqz") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "kube-api-access-6bxqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.761789 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93dbb123-819d-44f7-80f5-8c38c639f5f4-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.765279 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.791338 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-config-data" (OuterVolumeSpecName: "config-data") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.829430 4675 generic.go:334] "Generic (PLEG): container finished" podID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerID="f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e" exitCode=0 Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.829746 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.829768 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93dbb123-819d-44f7-80f5-8c38c639f5f4","Type":"ContainerDied","Data":"f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e"} Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.830866 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93dbb123-819d-44f7-80f5-8c38c639f5f4","Type":"ContainerDied","Data":"b38e71e5de19bb8f73b6cbd485738542f19363ab0d2195391d0b2a32a3a04cb8"} Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.830939 4675 scope.go:117] "RemoveContainer" containerID="f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850097 4675 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850137 4675 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93dbb123-819d-44f7-80f5-8c38c639f5f4-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850149 4675 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850157 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850168 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850179 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bxqz\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-kube-api-access-6bxqz\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850188 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850196 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.850204 4675 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93dbb123-819d-44f7-80f5-8c38c639f5f4-pod-info\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.876748 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-server-conf" (OuterVolumeSpecName: "server-conf") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.882902 4675 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.902591 4675 scope.go:117] "RemoveContainer" containerID="897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.952152 4675 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.952186 4675 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93dbb123-819d-44f7-80f5-8c38c639f5f4-server-conf\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.970590 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "93dbb123-819d-44f7-80f5-8c38c639f5f4" (UID: "93dbb123-819d-44f7-80f5-8c38c639f5f4"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.975317 4675 scope.go:117] "RemoveContainer" containerID="f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e" Feb 19 09:05:23 crc kubenswrapper[4675]: E0219 09:05:23.976196 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e\": container with ID starting with f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e not found: ID does not exist" containerID="f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.976275 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e"} err="failed to get container status \"f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e\": rpc error: code = NotFound desc = could not find container \"f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e\": container with ID starting with f10226f040e970785136118e7906bfa978355e4d15436585747911165885815e not found: ID does not exist" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.976308 4675 scope.go:117] "RemoveContainer" containerID="897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b" Feb 19 09:05:23 crc kubenswrapper[4675]: E0219 09:05:23.976738 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b\": container with ID starting with 897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b not found: ID does not exist" containerID="897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b" Feb 19 09:05:23 crc kubenswrapper[4675]: I0219 09:05:23.976776 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b"} err="failed to get container status \"897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b\": rpc error: code = NotFound desc = could not find container \"897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b\": container with ID starting with 897a0eb7ee27363d5f4158d9b9971cffe79039d335f36d77e4015c77547df42b not found: ID does not exist" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.053550 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93dbb123-819d-44f7-80f5-8c38c639f5f4-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.166163 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.177073 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.205890 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 09:05:24 crc kubenswrapper[4675]: E0219 09:05:24.206314 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerName="setup-container" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.206331 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerName="setup-container" Feb 19 09:05:24 crc kubenswrapper[4675]: E0219 09:05:24.206369 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerName="rabbitmq" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.206376 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerName="rabbitmq" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.206566 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" containerName="rabbitmq" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.207704 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.215547 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.215752 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.220666 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.220699 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.221179 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-f4b2m" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.221183 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.221281 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.236493 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.360820 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.361178 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.361312 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.361517 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.361572 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.361777 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-server-conf\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.361983 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.362037 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60253071-539f-43bb-a335-b351dbac4b77-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.362077 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-config-data\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.362161 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60253071-539f-43bb-a335-b351dbac4b77-pod-info\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.362191 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxrx8\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-kube-api-access-cxrx8\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464378 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464445 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464463 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464508 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-server-conf\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464535 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464554 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60253071-539f-43bb-a335-b351dbac4b77-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464581 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-config-data\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464625 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60253071-539f-43bb-a335-b351dbac4b77-pod-info\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464659 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxrx8\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-kube-api-access-cxrx8\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464691 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464739 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.464976 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.465092 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.465728 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.465834 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-config-data\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.465755 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.466484 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60253071-539f-43bb-a335-b351dbac4b77-server-conf\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.469951 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.470095 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.470439 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60253071-539f-43bb-a335-b351dbac4b77-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.471120 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60253071-539f-43bb-a335-b351dbac4b77-pod-info\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.488323 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxrx8\" (UniqueName: \"kubernetes.io/projected/60253071-539f-43bb-a335-b351dbac4b77-kube-api-access-cxrx8\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.501780 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"60253071-539f-43bb-a335-b351dbac4b77\") " pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.593438 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.844550 4675 generic.go:334] "Generic (PLEG): container finished" podID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerID="db61585170f749410d45a8cfd2979a86557f0fc24ab094b737b4fa31e9bf3613" exitCode=0 Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.844914 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1adde38f-79d6-4ee4-82ed-97004ffb7f85","Type":"ContainerDied","Data":"db61585170f749410d45a8cfd2979a86557f0fc24ab094b737b4fa31e9bf3613"} Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.857496 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985093 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-confd\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985157 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-tls\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985283 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-erlang-cookie\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985315 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985400 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1adde38f-79d6-4ee4-82ed-97004ffb7f85-erlang-cookie-secret\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985487 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-server-conf\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985576 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-plugins-conf\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985691 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1adde38f-79d6-4ee4-82ed-97004ffb7f85-pod-info\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985723 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rnc5\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-kube-api-access-9rnc5\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985753 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-config-data\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.985821 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-plugins\") pod \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\" (UID: \"1adde38f-79d6-4ee4-82ed-97004ffb7f85\") " Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.986091 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.986567 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.986632 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.987395 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:24 crc kubenswrapper[4675]: I0219 09:05:24.996903 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.008873 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1adde38f-79d6-4ee4-82ed-97004ffb7f85-pod-info" (OuterVolumeSpecName: "pod-info") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.009004 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1adde38f-79d6-4ee4-82ed-97004ffb7f85-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.018485 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.027336 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-kube-api-access-9rnc5" (OuterVolumeSpecName: "kube-api-access-9rnc5") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "kube-api-access-9rnc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.085274 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-config-data" (OuterVolumeSpecName: "config-data") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090394 4675 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090434 4675 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1adde38f-79d6-4ee4-82ed-97004ffb7f85-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090444 4675 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090454 4675 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1adde38f-79d6-4ee4-82ed-97004ffb7f85-pod-info\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090462 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rnc5\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-kube-api-access-9rnc5\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090472 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090480 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.090489 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.101944 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-server-conf" (OuterVolumeSpecName: "server-conf") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.114813 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93dbb123-819d-44f7-80f5-8c38c639f5f4" path="/var/lib/kubelet/pods/93dbb123-819d-44f7-80f5-8c38c639f5f4/volumes" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.131977 4675 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.188667 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.192898 4675 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.195289 4675 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1adde38f-79d6-4ee4-82ed-97004ffb7f85-server-conf\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.285796 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1adde38f-79d6-4ee4-82ed-97004ffb7f85" (UID: "1adde38f-79d6-4ee4-82ed-97004ffb7f85"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.298449 4675 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1adde38f-79d6-4ee4-82ed-97004ffb7f85-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.862496 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.862458 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1adde38f-79d6-4ee4-82ed-97004ffb7f85","Type":"ContainerDied","Data":"5c1fd77cea3aa2e47cce7e46f2baf1ed9a93d610455260f098319db176372cf4"} Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.862826 4675 scope.go:117] "RemoveContainer" containerID="db61585170f749410d45a8cfd2979a86557f0fc24ab094b737b4fa31e9bf3613" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.864688 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"60253071-539f-43bb-a335-b351dbac4b77","Type":"ContainerStarted","Data":"3904153cd4fe24d1c0bb4a1afa98dbb596ae6d5906d043026bfb9cd68b96bf7a"} Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.897729 4675 scope.go:117] "RemoveContainer" containerID="a993f5ff21a91f8f3b65ffb590d4332551df482a908b1a402992186078332313" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.937233 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.948565 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.984245 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 09:05:25 crc kubenswrapper[4675]: E0219 09:05:25.984685 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerName="rabbitmq" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.984704 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerName="rabbitmq" Feb 19 09:05:25 crc kubenswrapper[4675]: E0219 09:05:25.984730 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerName="setup-container" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.984736 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerName="setup-container" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.984907 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" containerName="rabbitmq" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.986062 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.990322 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.990847 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-88zj9" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.991057 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.991180 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.991183 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.991533 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 19 09:05:25 crc kubenswrapper[4675]: I0219 09:05:25.991629 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.007688 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.117340 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.117407 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.117435 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.117608 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.117728 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.117897 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.117985 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xcnv\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-kube-api-access-2xcnv\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.118026 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.118123 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.118171 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.118204 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.220135 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.220573 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.220611 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.221484 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.221689 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.221839 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.221915 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.221981 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222046 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222088 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222125 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222153 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222234 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222283 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xcnv\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-kube-api-access-2xcnv\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222294 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.222590 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.223084 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.277691 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.277828 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.278045 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.288873 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.289192 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xcnv\" (UniqueName: \"kubernetes.io/projected/94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4-kube-api-access-2xcnv\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.526249 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4\") " pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.615105 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.954720 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-zlnb9"] Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.956705 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.960355 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 19 09:05:26 crc kubenswrapper[4675]: I0219 09:05:26.971821 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-zlnb9"] Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.124822 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1adde38f-79d6-4ee4-82ed-97004ffb7f85" path="/var/lib/kubelet/pods/1adde38f-79d6-4ee4-82ed-97004ffb7f85/volumes" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.141143 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-config\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.141202 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.141254 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mv7s\" (UniqueName: \"kubernetes.io/projected/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-kube-api-access-2mv7s\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.141273 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.141299 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.141327 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-svc\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.141383 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.243254 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.243367 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-config\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.243391 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.243463 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mv7s\" (UniqueName: \"kubernetes.io/projected/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-kube-api-access-2mv7s\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.243485 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.243507 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.243541 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-svc\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.244891 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-config\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.245052 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.245253 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.245550 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.245746 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-svc\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.246294 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.272662 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mv7s\" (UniqueName: \"kubernetes.io/projected/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-kube-api-access-2mv7s\") pod \"dnsmasq-dns-5576978c7c-zlnb9\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.292516 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.327806 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.799457 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-zlnb9"] Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.892419 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"60253071-539f-43bb-a335-b351dbac4b77","Type":"ContainerStarted","Data":"2ddce43166a376d82173b3346f7e01036e4703851066af6d3c8ca89aaf5c1d49"} Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.893870 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4","Type":"ContainerStarted","Data":"e04f5b6b9204e59a8824c1e9dbf855fc70f80567c20312ec04a6d5524715f920"} Feb 19 09:05:27 crc kubenswrapper[4675]: I0219 09:05:27.895761 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" event={"ID":"6f9cbd5d-1cb1-48d2-89cb-e590984377a1","Type":"ContainerStarted","Data":"c4ef505c17177e953874947cbd50beecb2f14b9ac8755af184a87bdd6988437a"} Feb 19 09:05:28 crc kubenswrapper[4675]: I0219 09:05:28.907629 4675 generic.go:334] "Generic (PLEG): container finished" podID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerID="593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509" exitCode=0 Feb 19 09:05:28 crc kubenswrapper[4675]: I0219 09:05:28.907800 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" event={"ID":"6f9cbd5d-1cb1-48d2-89cb-e590984377a1","Type":"ContainerDied","Data":"593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509"} Feb 19 09:05:29 crc kubenswrapper[4675]: I0219 09:05:29.921388 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" event={"ID":"6f9cbd5d-1cb1-48d2-89cb-e590984377a1","Type":"ContainerStarted","Data":"bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c"} Feb 19 09:05:29 crc kubenswrapper[4675]: I0219 09:05:29.922818 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:29 crc kubenswrapper[4675]: I0219 09:05:29.923619 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4","Type":"ContainerStarted","Data":"0fa50e25032534fef78ae36231a13d1190c92005df2a7e5801d846efbf8f2ff3"} Feb 19 09:05:29 crc kubenswrapper[4675]: I0219 09:05:29.949979 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" podStartSLOduration=3.94996087 podStartE2EDuration="3.94996087s" podCreationTimestamp="2026-02-19 09:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:05:29.94321942 +0000 UTC m=+1311.570309698" watchObservedRunningTime="2026-02-19 09:05:29.94996087 +0000 UTC m=+1311.577051128" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.294541 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.353698 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-bn56v"] Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.353956 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" podUID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerName="dnsmasq-dns" containerID="cri-o://89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d" gracePeriod=10 Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.652185 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-8grnp"] Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.656725 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.673916 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-8grnp"] Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.773367 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.773462 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.773491 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m5dj\" (UniqueName: \"kubernetes.io/projected/012e6c07-a778-4850-99e0-47346310663b-kube-api-access-9m5dj\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.773903 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.774100 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.774238 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.774275 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-config\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.876973 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.877077 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.877144 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.877165 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-config\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.877234 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.877293 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m5dj\" (UniqueName: \"kubernetes.io/projected/012e6c07-a778-4850-99e0-47346310663b-kube-api-access-9m5dj\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.877312 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.878334 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.878946 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.879600 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-config\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.880248 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.880443 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.880458 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012e6c07-a778-4850-99e0-47346310663b-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.886827 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.908853 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m5dj\" (UniqueName: \"kubernetes.io/projected/012e6c07-a778-4850-99e0-47346310663b-kube-api-access-9m5dj\") pod \"dnsmasq-dns-8c6f6df99-8grnp\" (UID: \"012e6c07-a778-4850-99e0-47346310663b\") " pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.990943 4675 generic.go:334] "Generic (PLEG): container finished" podID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerID="89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d" exitCode=0 Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.990985 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" event={"ID":"634e48c1-5036-4a11-863d-b66f1c4fb341","Type":"ContainerDied","Data":"89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d"} Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.991012 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" event={"ID":"634e48c1-5036-4a11-863d-b66f1c4fb341","Type":"ContainerDied","Data":"135f13e11f7e61b51ada72a1180847e5aeb36a92849fbb7b255fec087b19113a"} Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.991028 4675 scope.go:117] "RemoveContainer" containerID="89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.991101 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-bn56v" Feb 19 09:05:37 crc kubenswrapper[4675]: I0219 09:05:37.998895 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.027868 4675 scope.go:117] "RemoveContainer" containerID="77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.079428 4675 scope.go:117] "RemoveContainer" containerID="89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.080317 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-sb\") pod \"634e48c1-5036-4a11-863d-b66f1c4fb341\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.080495 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-svc\") pod \"634e48c1-5036-4a11-863d-b66f1c4fb341\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.080953 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-config\") pod \"634e48c1-5036-4a11-863d-b66f1c4fb341\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.081028 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-nb\") pod \"634e48c1-5036-4a11-863d-b66f1c4fb341\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.081124 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzscp\" (UniqueName: \"kubernetes.io/projected/634e48c1-5036-4a11-863d-b66f1c4fb341-kube-api-access-qzscp\") pod \"634e48c1-5036-4a11-863d-b66f1c4fb341\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.081253 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-swift-storage-0\") pod \"634e48c1-5036-4a11-863d-b66f1c4fb341\" (UID: \"634e48c1-5036-4a11-863d-b66f1c4fb341\") " Feb 19 09:05:38 crc kubenswrapper[4675]: E0219 09:05:38.082524 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d\": container with ID starting with 89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d not found: ID does not exist" containerID="89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.082636 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d"} err="failed to get container status \"89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d\": rpc error: code = NotFound desc = could not find container \"89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d\": container with ID starting with 89f94d4aaa406d3b9ea5618e5afefe599c5497f948f90f0af6ced3cdb8fd2b5d not found: ID does not exist" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.082699 4675 scope.go:117] "RemoveContainer" containerID="77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a" Feb 19 09:05:38 crc kubenswrapper[4675]: E0219 09:05:38.090767 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a\": container with ID starting with 77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a not found: ID does not exist" containerID="77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.091521 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a"} err="failed to get container status \"77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a\": rpc error: code = NotFound desc = could not find container \"77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a\": container with ID starting with 77c32848f47f48d5f61bec21d6a5e055293505501fd7376ac9b763bf45c6dd1a not found: ID does not exist" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.108817 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/634e48c1-5036-4a11-863d-b66f1c4fb341-kube-api-access-qzscp" (OuterVolumeSpecName: "kube-api-access-qzscp") pod "634e48c1-5036-4a11-863d-b66f1c4fb341" (UID: "634e48c1-5036-4a11-863d-b66f1c4fb341"). InnerVolumeSpecName "kube-api-access-qzscp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.149075 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "634e48c1-5036-4a11-863d-b66f1c4fb341" (UID: "634e48c1-5036-4a11-863d-b66f1c4fb341"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.163830 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "634e48c1-5036-4a11-863d-b66f1c4fb341" (UID: "634e48c1-5036-4a11-863d-b66f1c4fb341"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.164863 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "634e48c1-5036-4a11-863d-b66f1c4fb341" (UID: "634e48c1-5036-4a11-863d-b66f1c4fb341"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.164899 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "634e48c1-5036-4a11-863d-b66f1c4fb341" (UID: "634e48c1-5036-4a11-863d-b66f1c4fb341"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.172759 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-config" (OuterVolumeSpecName: "config") pod "634e48c1-5036-4a11-863d-b66f1c4fb341" (UID: "634e48c1-5036-4a11-863d-b66f1c4fb341"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.185572 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.185615 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.185628 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.185661 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.185671 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/634e48c1-5036-4a11-863d-b66f1c4fb341-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.185679 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzscp\" (UniqueName: \"kubernetes.io/projected/634e48c1-5036-4a11-863d-b66f1c4fb341-kube-api-access-qzscp\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.329756 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-bn56v"] Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.337626 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-bn56v"] Feb 19 09:05:38 crc kubenswrapper[4675]: I0219 09:05:38.514897 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-8grnp"] Feb 19 09:05:39 crc kubenswrapper[4675]: I0219 09:05:39.015212 4675 generic.go:334] "Generic (PLEG): container finished" podID="012e6c07-a778-4850-99e0-47346310663b" containerID="5b17813f8f39e32bd3c41692752fcc4d52587f3df3e168396577c2fdb3b4b664" exitCode=0 Feb 19 09:05:39 crc kubenswrapper[4675]: I0219 09:05:39.015270 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" event={"ID":"012e6c07-a778-4850-99e0-47346310663b","Type":"ContainerDied","Data":"5b17813f8f39e32bd3c41692752fcc4d52587f3df3e168396577c2fdb3b4b664"} Feb 19 09:05:39 crc kubenswrapper[4675]: I0219 09:05:39.015304 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" event={"ID":"012e6c07-a778-4850-99e0-47346310663b","Type":"ContainerStarted","Data":"57955d262aa3b35b068f514823e692563b8ecdefea0f794d5061eeed17ee752a"} Feb 19 09:05:39 crc kubenswrapper[4675]: I0219 09:05:39.122103 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="634e48c1-5036-4a11-863d-b66f1c4fb341" path="/var/lib/kubelet/pods/634e48c1-5036-4a11-863d-b66f1c4fb341/volumes" Feb 19 09:05:40 crc kubenswrapper[4675]: I0219 09:05:40.029231 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" event={"ID":"012e6c07-a778-4850-99e0-47346310663b","Type":"ContainerStarted","Data":"44eda205364b4dbed0c03dccbb4aeaabc32e090317bc2509e83a6475e0acbd6c"} Feb 19 09:05:40 crc kubenswrapper[4675]: I0219 09:05:40.029520 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:40 crc kubenswrapper[4675]: I0219 09:05:40.052149 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" podStartSLOduration=3.052115415 podStartE2EDuration="3.052115415s" podCreationTimestamp="2026-02-19 09:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:05:40.048221572 +0000 UTC m=+1321.675311840" watchObservedRunningTime="2026-02-19 09:05:40.052115415 +0000 UTC m=+1321.679205723" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:47.999832 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-8grnp" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.065218 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-zlnb9"] Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.065511 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" podUID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerName="dnsmasq-dns" containerID="cri-o://bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c" gracePeriod=10 Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.570293 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.748116 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-config\") pod \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.748174 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mv7s\" (UniqueName: \"kubernetes.io/projected/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-kube-api-access-2mv7s\") pod \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.748225 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-nb\") pod \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.748312 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-openstack-edpm-ipam\") pod \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.748399 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-svc\") pod \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.748430 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-sb\") pod \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.748449 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-swift-storage-0\") pod \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\" (UID: \"6f9cbd5d-1cb1-48d2-89cb-e590984377a1\") " Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.756143 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-kube-api-access-2mv7s" (OuterVolumeSpecName: "kube-api-access-2mv7s") pod "6f9cbd5d-1cb1-48d2-89cb-e590984377a1" (UID: "6f9cbd5d-1cb1-48d2-89cb-e590984377a1"). InnerVolumeSpecName "kube-api-access-2mv7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.797736 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6f9cbd5d-1cb1-48d2-89cb-e590984377a1" (UID: "6f9cbd5d-1cb1-48d2-89cb-e590984377a1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.800015 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "6f9cbd5d-1cb1-48d2-89cb-e590984377a1" (UID: "6f9cbd5d-1cb1-48d2-89cb-e590984377a1"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.800240 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6f9cbd5d-1cb1-48d2-89cb-e590984377a1" (UID: "6f9cbd5d-1cb1-48d2-89cb-e590984377a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.802905 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f9cbd5d-1cb1-48d2-89cb-e590984377a1" (UID: "6f9cbd5d-1cb1-48d2-89cb-e590984377a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.808146 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-config" (OuterVolumeSpecName: "config") pod "6f9cbd5d-1cb1-48d2-89cb-e590984377a1" (UID: "6f9cbd5d-1cb1-48d2-89cb-e590984377a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.810170 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f9cbd5d-1cb1-48d2-89cb-e590984377a1" (UID: "6f9cbd5d-1cb1-48d2-89cb-e590984377a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.851234 4675 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.851272 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mv7s\" (UniqueName: \"kubernetes.io/projected/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-kube-api-access-2mv7s\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.851284 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.851294 4675 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.851303 4675 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.851315 4675 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:48 crc kubenswrapper[4675]: I0219 09:05:48.851326 4675 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f9cbd5d-1cb1-48d2-89cb-e590984377a1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.122910 4675 generic.go:334] "Generic (PLEG): container finished" podID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerID="bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c" exitCode=0 Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.122960 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" event={"ID":"6f9cbd5d-1cb1-48d2-89cb-e590984377a1","Type":"ContainerDied","Data":"bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c"} Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.123269 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" event={"ID":"6f9cbd5d-1cb1-48d2-89cb-e590984377a1","Type":"ContainerDied","Data":"c4ef505c17177e953874947cbd50beecb2f14b9ac8755af184a87bdd6988437a"} Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.123298 4675 scope.go:117] "RemoveContainer" containerID="bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c" Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.122990 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-zlnb9" Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.153878 4675 scope.go:117] "RemoveContainer" containerID="593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509" Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.158025 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-zlnb9"] Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.168414 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-zlnb9"] Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.177182 4675 scope.go:117] "RemoveContainer" containerID="bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c" Feb 19 09:05:49 crc kubenswrapper[4675]: E0219 09:05:49.177723 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c\": container with ID starting with bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c not found: ID does not exist" containerID="bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c" Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.177765 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c"} err="failed to get container status \"bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c\": rpc error: code = NotFound desc = could not find container \"bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c\": container with ID starting with bc56a75589ecfe7c8159534d6f26e3b4a35f245c53b83fae7ffc73a96336214c not found: ID does not exist" Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.177793 4675 scope.go:117] "RemoveContainer" containerID="593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509" Feb 19 09:05:49 crc kubenswrapper[4675]: E0219 09:05:49.178183 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509\": container with ID starting with 593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509 not found: ID does not exist" containerID="593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509" Feb 19 09:05:49 crc kubenswrapper[4675]: I0219 09:05:49.178235 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509"} err="failed to get container status \"593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509\": rpc error: code = NotFound desc = could not find container \"593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509\": container with ID starting with 593925c6f471d681cecd5146600ee97893dd422b71dbe2e9e549ddaee3adf509 not found: ID does not exist" Feb 19 09:05:51 crc kubenswrapper[4675]: I0219 09:05:51.115162 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" path="/var/lib/kubelet/pods/6f9cbd5d-1cb1-48d2-89cb-e590984377a1/volumes" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.221706 4675 generic.go:334] "Generic (PLEG): container finished" podID="60253071-539f-43bb-a335-b351dbac4b77" containerID="2ddce43166a376d82173b3346f7e01036e4703851066af6d3c8ca89aaf5c1d49" exitCode=0 Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.221788 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"60253071-539f-43bb-a335-b351dbac4b77","Type":"ContainerDied","Data":"2ddce43166a376d82173b3346f7e01036e4703851066af6d3c8ca89aaf5c1d49"} Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.925159 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp"] Feb 19 09:06:00 crc kubenswrapper[4675]: E0219 09:06:00.925943 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerName="dnsmasq-dns" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.925964 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerName="dnsmasq-dns" Feb 19 09:06:00 crc kubenswrapper[4675]: E0219 09:06:00.925984 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerName="dnsmasq-dns" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.925990 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerName="dnsmasq-dns" Feb 19 09:06:00 crc kubenswrapper[4675]: E0219 09:06:00.925999 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerName="init" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.926005 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerName="init" Feb 19 09:06:00 crc kubenswrapper[4675]: E0219 09:06:00.926027 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerName="init" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.926033 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerName="init" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.926254 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f9cbd5d-1cb1-48d2-89cb-e590984377a1" containerName="dnsmasq-dns" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.926270 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="634e48c1-5036-4a11-863d-b66f1c4fb341" containerName="dnsmasq-dns" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.926864 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.930758 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.931180 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.931308 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.931399 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:06:00 crc kubenswrapper[4675]: I0219 09:06:00.959754 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp"] Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.094681 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.094758 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.094867 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk4k2\" (UniqueName: \"kubernetes.io/projected/34f8029c-b96e-42c2-8005-ebc9988389b7-kube-api-access-sk4k2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.094967 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.197014 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.197515 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.197680 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.198311 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk4k2\" (UniqueName: \"kubernetes.io/projected/34f8029c-b96e-42c2-8005-ebc9988389b7-kube-api-access-sk4k2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.204035 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.204829 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.207211 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.218431 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk4k2\" (UniqueName: \"kubernetes.io/projected/34f8029c-b96e-42c2-8005-ebc9988389b7-kube-api-access-sk4k2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.239813 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"60253071-539f-43bb-a335-b351dbac4b77","Type":"ContainerStarted","Data":"031f8a9776c960305aeaf478ffc08b02224efcd6027ae0622a012d52ee8e5e09"} Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.240104 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.261868 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.276791 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.276741908 podStartE2EDuration="37.276741908s" podCreationTimestamp="2026-02-19 09:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:06:01.266501625 +0000 UTC m=+1342.893591893" watchObservedRunningTime="2026-02-19 09:06:01.276741908 +0000 UTC m=+1342.903832176" Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.830401 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp"] Feb 19 09:06:01 crc kubenswrapper[4675]: W0219 09:06:01.831048 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34f8029c_b96e_42c2_8005_ebc9988389b7.slice/crio-5c56c2bd643666555e1fe3ceca918a5554dddba5a2cfed8f3954e9864e57ba03 WatchSource:0}: Error finding container 5c56c2bd643666555e1fe3ceca918a5554dddba5a2cfed8f3954e9864e57ba03: Status 404 returned error can't find the container with id 5c56c2bd643666555e1fe3ceca918a5554dddba5a2cfed8f3954e9864e57ba03 Feb 19 09:06:01 crc kubenswrapper[4675]: I0219 09:06:01.833562 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:06:02 crc kubenswrapper[4675]: I0219 09:06:02.249160 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" event={"ID":"34f8029c-b96e-42c2-8005-ebc9988389b7","Type":"ContainerStarted","Data":"5c56c2bd643666555e1fe3ceca918a5554dddba5a2cfed8f3954e9864e57ba03"} Feb 19 09:06:02 crc kubenswrapper[4675]: I0219 09:06:02.251321 4675 generic.go:334] "Generic (PLEG): container finished" podID="94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4" containerID="0fa50e25032534fef78ae36231a13d1190c92005df2a7e5801d846efbf8f2ff3" exitCode=0 Feb 19 09:06:02 crc kubenswrapper[4675]: I0219 09:06:02.251412 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4","Type":"ContainerDied","Data":"0fa50e25032534fef78ae36231a13d1190c92005df2a7e5801d846efbf8f2ff3"} Feb 19 09:06:03 crc kubenswrapper[4675]: I0219 09:06:03.281239 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4","Type":"ContainerStarted","Data":"9ef9efaa1e4c828b5e570badd8aaf999d0a8ae84c0e69a404355f27c801152c8"} Feb 19 09:06:03 crc kubenswrapper[4675]: I0219 09:06:03.281785 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:06:03 crc kubenswrapper[4675]: I0219 09:06:03.317418 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.317399206 podStartE2EDuration="38.317399206s" podCreationTimestamp="2026-02-19 09:05:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:06:03.31004434 +0000 UTC m=+1344.937134628" watchObservedRunningTime="2026-02-19 09:06:03.317399206 +0000 UTC m=+1344.944489474" Feb 19 09:06:12 crc kubenswrapper[4675]: I0219 09:06:12.395863 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" event={"ID":"34f8029c-b96e-42c2-8005-ebc9988389b7","Type":"ContainerStarted","Data":"1738fafa1c586bc72e14a48faae4957f12ea3c4b2dbebb3005c8c0c5da924977"} Feb 19 09:06:12 crc kubenswrapper[4675]: I0219 09:06:12.420255 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" podStartSLOduration=2.637669492 podStartE2EDuration="12.42023635s" podCreationTimestamp="2026-02-19 09:06:00 +0000 UTC" firstStartedPulling="2026-02-19 09:06:01.833338839 +0000 UTC m=+1343.460429107" lastFinishedPulling="2026-02-19 09:06:11.615905697 +0000 UTC m=+1353.242995965" observedRunningTime="2026-02-19 09:06:12.415938516 +0000 UTC m=+1354.043028784" watchObservedRunningTime="2026-02-19 09:06:12.42023635 +0000 UTC m=+1354.047326608" Feb 19 09:06:14 crc kubenswrapper[4675]: I0219 09:06:14.597903 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 19 09:06:16 crc kubenswrapper[4675]: I0219 09:06:16.617871 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 19 09:06:22 crc kubenswrapper[4675]: I0219 09:06:22.480494 4675 generic.go:334] "Generic (PLEG): container finished" podID="34f8029c-b96e-42c2-8005-ebc9988389b7" containerID="1738fafa1c586bc72e14a48faae4957f12ea3c4b2dbebb3005c8c0c5da924977" exitCode=0 Feb 19 09:06:22 crc kubenswrapper[4675]: I0219 09:06:22.480542 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" event={"ID":"34f8029c-b96e-42c2-8005-ebc9988389b7","Type":"ContainerDied","Data":"1738fafa1c586bc72e14a48faae4957f12ea3c4b2dbebb3005c8c0c5da924977"} Feb 19 09:06:23 crc kubenswrapper[4675]: I0219 09:06:23.931923 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.045685 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk4k2\" (UniqueName: \"kubernetes.io/projected/34f8029c-b96e-42c2-8005-ebc9988389b7-kube-api-access-sk4k2\") pod \"34f8029c-b96e-42c2-8005-ebc9988389b7\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.045734 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-repo-setup-combined-ca-bundle\") pod \"34f8029c-b96e-42c2-8005-ebc9988389b7\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.045835 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-inventory\") pod \"34f8029c-b96e-42c2-8005-ebc9988389b7\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.045854 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-ssh-key-openstack-edpm-ipam\") pod \"34f8029c-b96e-42c2-8005-ebc9988389b7\" (UID: \"34f8029c-b96e-42c2-8005-ebc9988389b7\") " Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.056271 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "34f8029c-b96e-42c2-8005-ebc9988389b7" (UID: "34f8029c-b96e-42c2-8005-ebc9988389b7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.056664 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f8029c-b96e-42c2-8005-ebc9988389b7-kube-api-access-sk4k2" (OuterVolumeSpecName: "kube-api-access-sk4k2") pod "34f8029c-b96e-42c2-8005-ebc9988389b7" (UID: "34f8029c-b96e-42c2-8005-ebc9988389b7"). InnerVolumeSpecName "kube-api-access-sk4k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.075786 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-inventory" (OuterVolumeSpecName: "inventory") pod "34f8029c-b96e-42c2-8005-ebc9988389b7" (UID: "34f8029c-b96e-42c2-8005-ebc9988389b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.076884 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "34f8029c-b96e-42c2-8005-ebc9988389b7" (UID: "34f8029c-b96e-42c2-8005-ebc9988389b7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.148523 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.148575 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk4k2\" (UniqueName: \"kubernetes.io/projected/34f8029c-b96e-42c2-8005-ebc9988389b7-kube-api-access-sk4k2\") on node \"crc\" DevicePath \"\"" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.148591 4675 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.148605 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34f8029c-b96e-42c2-8005-ebc9988389b7-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.498509 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" event={"ID":"34f8029c-b96e-42c2-8005-ebc9988389b7","Type":"ContainerDied","Data":"5c56c2bd643666555e1fe3ceca918a5554dddba5a2cfed8f3954e9864e57ba03"} Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.498564 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.498569 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c56c2bd643666555e1fe3ceca918a5554dddba5a2cfed8f3954e9864e57ba03" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.569805 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8"] Feb 19 09:06:24 crc kubenswrapper[4675]: E0219 09:06:24.570316 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f8029c-b96e-42c2-8005-ebc9988389b7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.570341 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f8029c-b96e-42c2-8005-ebc9988389b7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.570613 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f8029c-b96e-42c2-8005-ebc9988389b7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.571522 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.574599 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.574836 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.574864 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.576283 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.580090 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8"] Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.658631 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.658933 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.659062 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwnbl\" (UniqueName: \"kubernetes.io/projected/185defec-09d9-4992-bc76-fc2cc62a788e-kube-api-access-bwnbl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.760988 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.761098 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.761155 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwnbl\" (UniqueName: \"kubernetes.io/projected/185defec-09d9-4992-bc76-fc2cc62a788e-kube-api-access-bwnbl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.766705 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.767464 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.776786 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwnbl\" (UniqueName: \"kubernetes.io/projected/185defec-09d9-4992-bc76-fc2cc62a788e-kube-api-access-bwnbl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zkwd8\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:24 crc kubenswrapper[4675]: I0219 09:06:24.898631 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:25 crc kubenswrapper[4675]: I0219 09:06:25.448760 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8"] Feb 19 09:06:25 crc kubenswrapper[4675]: I0219 09:06:25.529875 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" event={"ID":"185defec-09d9-4992-bc76-fc2cc62a788e","Type":"ContainerStarted","Data":"e8a8d8c533cad535c686cae8c6dee8ceb6ebfaa1286669c8aa0ed68d700d76a3"} Feb 19 09:06:26 crc kubenswrapper[4675]: I0219 09:06:26.543293 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" event={"ID":"185defec-09d9-4992-bc76-fc2cc62a788e","Type":"ContainerStarted","Data":"c576f4c06ab4445f2cb63c542a99b47b55274b6262df5fe4efc07a7bf27b16aa"} Feb 19 09:06:26 crc kubenswrapper[4675]: I0219 09:06:26.570558 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" podStartSLOduration=2.084582183 podStartE2EDuration="2.570530933s" podCreationTimestamp="2026-02-19 09:06:24 +0000 UTC" firstStartedPulling="2026-02-19 09:06:25.45740283 +0000 UTC m=+1367.084493098" lastFinishedPulling="2026-02-19 09:06:25.94335158 +0000 UTC m=+1367.570441848" observedRunningTime="2026-02-19 09:06:26.558113002 +0000 UTC m=+1368.185203270" watchObservedRunningTime="2026-02-19 09:06:26.570530933 +0000 UTC m=+1368.197621201" Feb 19 09:06:29 crc kubenswrapper[4675]: I0219 09:06:29.571719 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" event={"ID":"185defec-09d9-4992-bc76-fc2cc62a788e","Type":"ContainerDied","Data":"c576f4c06ab4445f2cb63c542a99b47b55274b6262df5fe4efc07a7bf27b16aa"} Feb 19 09:06:29 crc kubenswrapper[4675]: I0219 09:06:29.571626 4675 generic.go:334] "Generic (PLEG): container finished" podID="185defec-09d9-4992-bc76-fc2cc62a788e" containerID="c576f4c06ab4445f2cb63c542a99b47b55274b6262df5fe4efc07a7bf27b16aa" exitCode=0 Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.065156 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.125085 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-inventory\") pod \"185defec-09d9-4992-bc76-fc2cc62a788e\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.125263 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-ssh-key-openstack-edpm-ipam\") pod \"185defec-09d9-4992-bc76-fc2cc62a788e\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.125304 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwnbl\" (UniqueName: \"kubernetes.io/projected/185defec-09d9-4992-bc76-fc2cc62a788e-kube-api-access-bwnbl\") pod \"185defec-09d9-4992-bc76-fc2cc62a788e\" (UID: \"185defec-09d9-4992-bc76-fc2cc62a788e\") " Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.133305 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185defec-09d9-4992-bc76-fc2cc62a788e-kube-api-access-bwnbl" (OuterVolumeSpecName: "kube-api-access-bwnbl") pod "185defec-09d9-4992-bc76-fc2cc62a788e" (UID: "185defec-09d9-4992-bc76-fc2cc62a788e"). InnerVolumeSpecName "kube-api-access-bwnbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.157403 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "185defec-09d9-4992-bc76-fc2cc62a788e" (UID: "185defec-09d9-4992-bc76-fc2cc62a788e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.158992 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-inventory" (OuterVolumeSpecName: "inventory") pod "185defec-09d9-4992-bc76-fc2cc62a788e" (UID: "185defec-09d9-4992-bc76-fc2cc62a788e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.227457 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.227593 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185defec-09d9-4992-bc76-fc2cc62a788e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.228310 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwnbl\" (UniqueName: \"kubernetes.io/projected/185defec-09d9-4992-bc76-fc2cc62a788e-kube-api-access-bwnbl\") on node \"crc\" DevicePath \"\"" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.591679 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" event={"ID":"185defec-09d9-4992-bc76-fc2cc62a788e","Type":"ContainerDied","Data":"e8a8d8c533cad535c686cae8c6dee8ceb6ebfaa1286669c8aa0ed68d700d76a3"} Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.591718 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zkwd8" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.591734 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8a8d8c533cad535c686cae8c6dee8ceb6ebfaa1286669c8aa0ed68d700d76a3" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.671181 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd"] Feb 19 09:06:31 crc kubenswrapper[4675]: E0219 09:06:31.671631 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185defec-09d9-4992-bc76-fc2cc62a788e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.671653 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="185defec-09d9-4992-bc76-fc2cc62a788e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.671892 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="185defec-09d9-4992-bc76-fc2cc62a788e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.672565 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.677059 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.677761 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.677974 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.678105 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.691716 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd"] Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.737562 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.737651 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.737822 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.738000 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flgjq\" (UniqueName: \"kubernetes.io/projected/df5fab89-7323-474a-96b4-60dbd70e2c50-kube-api-access-flgjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.839770 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.840234 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.840371 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.840520 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flgjq\" (UniqueName: \"kubernetes.io/projected/df5fab89-7323-474a-96b4-60dbd70e2c50-kube-api-access-flgjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.846363 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.846603 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.849601 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:31 crc kubenswrapper[4675]: I0219 09:06:31.871066 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flgjq\" (UniqueName: \"kubernetes.io/projected/df5fab89-7323-474a-96b4-60dbd70e2c50-kube-api-access-flgjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:32 crc kubenswrapper[4675]: I0219 09:06:32.001248 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:06:32 crc kubenswrapper[4675]: I0219 09:06:32.550785 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd"] Feb 19 09:06:32 crc kubenswrapper[4675]: I0219 09:06:32.606498 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" event={"ID":"df5fab89-7323-474a-96b4-60dbd70e2c50","Type":"ContainerStarted","Data":"8eb28248ab864447b13b7e6fdede4a080b67cb2c386d80e816688af91cb92745"} Feb 19 09:06:33 crc kubenswrapper[4675]: I0219 09:06:33.618895 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" event={"ID":"df5fab89-7323-474a-96b4-60dbd70e2c50","Type":"ContainerStarted","Data":"29c91f583d491a81e3b421705ba4c40ab1d6b9a1f891569089d220d94d630770"} Feb 19 09:06:33 crc kubenswrapper[4675]: I0219 09:06:33.645321 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" podStartSLOduration=2.220155316 podStartE2EDuration="2.645297135s" podCreationTimestamp="2026-02-19 09:06:31 +0000 UTC" firstStartedPulling="2026-02-19 09:06:32.560889687 +0000 UTC m=+1374.187979955" lastFinishedPulling="2026-02-19 09:06:32.986031506 +0000 UTC m=+1374.613121774" observedRunningTime="2026-02-19 09:06:33.636986474 +0000 UTC m=+1375.264076752" watchObservedRunningTime="2026-02-19 09:06:33.645297135 +0000 UTC m=+1375.272387403" Feb 19 09:06:51 crc kubenswrapper[4675]: I0219 09:06:51.712380 4675 scope.go:117] "RemoveContainer" containerID="f452536d0b73e2b61a8adf745b19b101d7c5a43a6d7d8adb0eec3a4df9e9bdab" Feb 19 09:06:51 crc kubenswrapper[4675]: I0219 09:06:51.734396 4675 scope.go:117] "RemoveContainer" containerID="27e6b292808a508110eee774e6ba38c1d382dac63eda9b8635c1c8d7aa0f15b5" Feb 19 09:06:51 crc kubenswrapper[4675]: I0219 09:06:51.758341 4675 scope.go:117] "RemoveContainer" containerID="2d0218220dd3aef3e157051f3dabc4234121c059b6dc78682887daed25d83580" Feb 19 09:06:51 crc kubenswrapper[4675]: I0219 09:06:51.807692 4675 scope.go:117] "RemoveContainer" containerID="69153e694360781dae004fbfd2eeebd54afe3db2ba964ad120a8cabb0f562602" Feb 19 09:07:11 crc kubenswrapper[4675]: I0219 09:07:11.743944 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:07:11 crc kubenswrapper[4675]: I0219 09:07:11.744936 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:07:41 crc kubenswrapper[4675]: I0219 09:07:41.744051 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:07:41 crc kubenswrapper[4675]: I0219 09:07:41.745975 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:07:51 crc kubenswrapper[4675]: I0219 09:07:51.912408 4675 scope.go:117] "RemoveContainer" containerID="b801f0662ccb083ec945731718f55feff253d67d407722cdf02706cd5333269d" Feb 19 09:07:51 crc kubenswrapper[4675]: I0219 09:07:51.949652 4675 scope.go:117] "RemoveContainer" containerID="323d24ff370c4234b2883d3742bba1699b5a106f42206d66ef5b910abcdb95a1" Feb 19 09:07:52 crc kubenswrapper[4675]: I0219 09:07:52.031355 4675 scope.go:117] "RemoveContainer" containerID="322a7518cdb040945e0d8ac2593531619491766439e54623fbf6105fd05ea3b8" Feb 19 09:07:52 crc kubenswrapper[4675]: I0219 09:07:52.058031 4675 scope.go:117] "RemoveContainer" containerID="ea3815d81b1cf5ae40f7aec970ce31de56cb826043899ccd1394fda58b38191d" Feb 19 09:07:52 crc kubenswrapper[4675]: I0219 09:07:52.088596 4675 scope.go:117] "RemoveContainer" containerID="25f975d840cb9091c25aec3d46da36d20c77174a609fa0e249c0fc4c7be63e00" Feb 19 09:08:11 crc kubenswrapper[4675]: I0219 09:08:11.743566 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:08:11 crc kubenswrapper[4675]: I0219 09:08:11.744424 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:08:11 crc kubenswrapper[4675]: I0219 09:08:11.744478 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:08:11 crc kubenswrapper[4675]: I0219 09:08:11.745398 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d8cec9b5ad0c194b9e5ae1c687b8aae050e8cf77a0f404f058bb11195c4e24e3"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:08:11 crc kubenswrapper[4675]: I0219 09:08:11.745469 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://d8cec9b5ad0c194b9e5ae1c687b8aae050e8cf77a0f404f058bb11195c4e24e3" gracePeriod=600 Feb 19 09:08:12 crc kubenswrapper[4675]: I0219 09:08:12.705446 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="d8cec9b5ad0c194b9e5ae1c687b8aae050e8cf77a0f404f058bb11195c4e24e3" exitCode=0 Feb 19 09:08:12 crc kubenswrapper[4675]: I0219 09:08:12.706251 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"d8cec9b5ad0c194b9e5ae1c687b8aae050e8cf77a0f404f058bb11195c4e24e3"} Feb 19 09:08:12 crc kubenswrapper[4675]: I0219 09:08:12.706286 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44"} Feb 19 09:08:12 crc kubenswrapper[4675]: I0219 09:08:12.706324 4675 scope.go:117] "RemoveContainer" containerID="29c6cf62bac0dde894372357f7b07e5a365b8cc260e995aa58cda63d08dcff6f" Feb 19 09:08:52 crc kubenswrapper[4675]: I0219 09:08:52.227587 4675 scope.go:117] "RemoveContainer" containerID="c9173343cbea7deb828160fa7599a578a21600e26e0722f6d0dfaca422335b38" Feb 19 09:08:52 crc kubenswrapper[4675]: I0219 09:08:52.267332 4675 scope.go:117] "RemoveContainer" containerID="ec9226b305f0ce2cd0d2f932df6518705ee9d7c236b0bd4e5ad19b7f29b30918" Feb 19 09:09:32 crc kubenswrapper[4675]: I0219 09:09:32.495266 4675 generic.go:334] "Generic (PLEG): container finished" podID="df5fab89-7323-474a-96b4-60dbd70e2c50" containerID="29c91f583d491a81e3b421705ba4c40ab1d6b9a1f891569089d220d94d630770" exitCode=0 Feb 19 09:09:32 crc kubenswrapper[4675]: I0219 09:09:32.495358 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" event={"ID":"df5fab89-7323-474a-96b4-60dbd70e2c50","Type":"ContainerDied","Data":"29c91f583d491a81e3b421705ba4c40ab1d6b9a1f891569089d220d94d630770"} Feb 19 09:09:33 crc kubenswrapper[4675]: I0219 09:09:33.983739 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.014793 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-ssh-key-openstack-edpm-ipam\") pod \"df5fab89-7323-474a-96b4-60dbd70e2c50\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.014903 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flgjq\" (UniqueName: \"kubernetes.io/projected/df5fab89-7323-474a-96b4-60dbd70e2c50-kube-api-access-flgjq\") pod \"df5fab89-7323-474a-96b4-60dbd70e2c50\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.014975 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-inventory\") pod \"df5fab89-7323-474a-96b4-60dbd70e2c50\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.015107 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-bootstrap-combined-ca-bundle\") pod \"df5fab89-7323-474a-96b4-60dbd70e2c50\" (UID: \"df5fab89-7323-474a-96b4-60dbd70e2c50\") " Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.057438 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df5fab89-7323-474a-96b4-60dbd70e2c50-kube-api-access-flgjq" (OuterVolumeSpecName: "kube-api-access-flgjq") pod "df5fab89-7323-474a-96b4-60dbd70e2c50" (UID: "df5fab89-7323-474a-96b4-60dbd70e2c50"). InnerVolumeSpecName "kube-api-access-flgjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.057525 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "df5fab89-7323-474a-96b4-60dbd70e2c50" (UID: "df5fab89-7323-474a-96b4-60dbd70e2c50"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.067523 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "df5fab89-7323-474a-96b4-60dbd70e2c50" (UID: "df5fab89-7323-474a-96b4-60dbd70e2c50"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.067615 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-inventory" (OuterVolumeSpecName: "inventory") pod "df5fab89-7323-474a-96b4-60dbd70e2c50" (UID: "df5fab89-7323-474a-96b4-60dbd70e2c50"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.117472 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.117520 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flgjq\" (UniqueName: \"kubernetes.io/projected/df5fab89-7323-474a-96b4-60dbd70e2c50-kube-api-access-flgjq\") on node \"crc\" DevicePath \"\"" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.117534 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.117546 4675 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5fab89-7323-474a-96b4-60dbd70e2c50-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.517371 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" event={"ID":"df5fab89-7323-474a-96b4-60dbd70e2c50","Type":"ContainerDied","Data":"8eb28248ab864447b13b7e6fdede4a080b67cb2c386d80e816688af91cb92745"} Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.517420 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eb28248ab864447b13b7e6fdede4a080b67cb2c386d80e816688af91cb92745" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.518109 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.608600 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t"] Feb 19 09:09:34 crc kubenswrapper[4675]: E0219 09:09:34.609089 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5fab89-7323-474a-96b4-60dbd70e2c50" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.609110 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5fab89-7323-474a-96b4-60dbd70e2c50" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.609300 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="df5fab89-7323-474a-96b4-60dbd70e2c50" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.610001 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.612477 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.612769 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.612788 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.614844 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.622466 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t"] Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.628133 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.628350 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xgg6\" (UniqueName: \"kubernetes.io/projected/31e7b53a-04a2-4fd3-b18e-893da088a60f-kube-api-access-7xgg6\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.628491 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.729744 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xgg6\" (UniqueName: \"kubernetes.io/projected/31e7b53a-04a2-4fd3-b18e-893da088a60f-kube-api-access-7xgg6\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.729811 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.729841 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.736009 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.744751 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.751877 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xgg6\" (UniqueName: \"kubernetes.io/projected/31e7b53a-04a2-4fd3-b18e-893da088a60f-kube-api-access-7xgg6\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fr44t\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:34 crc kubenswrapper[4675]: I0219 09:09:34.928335 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:09:35 crc kubenswrapper[4675]: I0219 09:09:35.304307 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t"] Feb 19 09:09:35 crc kubenswrapper[4675]: W0219 09:09:35.315767 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31e7b53a_04a2_4fd3_b18e_893da088a60f.slice/crio-0eb6a3c9b9561341f6c99696e74c7ef6ec7a7143d90d8ce6d09bdb51487bf34e WatchSource:0}: Error finding container 0eb6a3c9b9561341f6c99696e74c7ef6ec7a7143d90d8ce6d09bdb51487bf34e: Status 404 returned error can't find the container with id 0eb6a3c9b9561341f6c99696e74c7ef6ec7a7143d90d8ce6d09bdb51487bf34e Feb 19 09:09:35 crc kubenswrapper[4675]: I0219 09:09:35.526713 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" event={"ID":"31e7b53a-04a2-4fd3-b18e-893da088a60f","Type":"ContainerStarted","Data":"0eb6a3c9b9561341f6c99696e74c7ef6ec7a7143d90d8ce6d09bdb51487bf34e"} Feb 19 09:09:36 crc kubenswrapper[4675]: I0219 09:09:36.538741 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" event={"ID":"31e7b53a-04a2-4fd3-b18e-893da088a60f","Type":"ContainerStarted","Data":"d1b13e2c2b5ce477faa4320a9eb9b3f0045c49fab4802c22767a1d29a0f5138e"} Feb 19 09:09:36 crc kubenswrapper[4675]: I0219 09:09:36.558365 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" podStartSLOduration=2.071542246 podStartE2EDuration="2.558342704s" podCreationTimestamp="2026-02-19 09:09:34 +0000 UTC" firstStartedPulling="2026-02-19 09:09:35.317669313 +0000 UTC m=+1556.944759581" lastFinishedPulling="2026-02-19 09:09:35.804469771 +0000 UTC m=+1557.431560039" observedRunningTime="2026-02-19 09:09:36.556241639 +0000 UTC m=+1558.183331927" watchObservedRunningTime="2026-02-19 09:09:36.558342704 +0000 UTC m=+1558.185432972" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.038432 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frdcj"] Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.041389 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.049189 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frdcj"] Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.176140 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-utilities\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.176221 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-catalog-content\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.176862 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjp7w\" (UniqueName: \"kubernetes.io/projected/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-kube-api-access-gjp7w\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.278471 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjp7w\" (UniqueName: \"kubernetes.io/projected/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-kube-api-access-gjp7w\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.278588 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-utilities\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.278611 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-catalog-content\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.279204 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-utilities\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.279588 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-catalog-content\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.300589 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjp7w\" (UniqueName: \"kubernetes.io/projected/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-kube-api-access-gjp7w\") pod \"redhat-operators-frdcj\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.376285 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:09:55 crc kubenswrapper[4675]: I0219 09:09:55.864593 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frdcj"] Feb 19 09:09:56 crc kubenswrapper[4675]: I0219 09:09:56.716545 4675 generic.go:334] "Generic (PLEG): container finished" podID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerID="1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762" exitCode=0 Feb 19 09:09:56 crc kubenswrapper[4675]: I0219 09:09:56.716979 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdcj" event={"ID":"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af","Type":"ContainerDied","Data":"1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762"} Feb 19 09:09:56 crc kubenswrapper[4675]: I0219 09:09:56.717007 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdcj" event={"ID":"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af","Type":"ContainerStarted","Data":"c8776503ac16f3b1282ea3e2961476343541cdc042151395171a1dc694350c3e"} Feb 19 09:09:58 crc kubenswrapper[4675]: I0219 09:09:58.736523 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdcj" event={"ID":"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af","Type":"ContainerStarted","Data":"e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2"} Feb 19 09:09:59 crc kubenswrapper[4675]: I0219 09:09:59.749732 4675 generic.go:334] "Generic (PLEG): container finished" podID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerID="e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2" exitCode=0 Feb 19 09:09:59 crc kubenswrapper[4675]: I0219 09:09:59.749791 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdcj" event={"ID":"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af","Type":"ContainerDied","Data":"e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2"} Feb 19 09:10:01 crc kubenswrapper[4675]: I0219 09:10:01.773112 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdcj" event={"ID":"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af","Type":"ContainerStarted","Data":"8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad"} Feb 19 09:10:01 crc kubenswrapper[4675]: I0219 09:10:01.794370 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frdcj" podStartSLOduration=3.348479071 podStartE2EDuration="6.794345763s" podCreationTimestamp="2026-02-19 09:09:55 +0000 UTC" firstStartedPulling="2026-02-19 09:09:56.718835606 +0000 UTC m=+1578.345925874" lastFinishedPulling="2026-02-19 09:10:00.164702298 +0000 UTC m=+1581.791792566" observedRunningTime="2026-02-19 09:10:01.789104096 +0000 UTC m=+1583.416194384" watchObservedRunningTime="2026-02-19 09:10:01.794345763 +0000 UTC m=+1583.421436031" Feb 19 09:10:05 crc kubenswrapper[4675]: I0219 09:10:05.376436 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:10:05 crc kubenswrapper[4675]: I0219 09:10:05.377835 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:10:06 crc kubenswrapper[4675]: I0219 09:10:06.431665 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frdcj" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="registry-server" probeResult="failure" output=< Feb 19 09:10:06 crc kubenswrapper[4675]: timeout: failed to connect service ":50051" within 1s Feb 19 09:10:06 crc kubenswrapper[4675]: > Feb 19 09:10:15 crc kubenswrapper[4675]: I0219 09:10:15.421586 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:10:15 crc kubenswrapper[4675]: I0219 09:10:15.472387 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:10:15 crc kubenswrapper[4675]: I0219 09:10:15.657409 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frdcj"] Feb 19 09:10:16 crc kubenswrapper[4675]: I0219 09:10:16.927986 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frdcj" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="registry-server" containerID="cri-o://8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad" gracePeriod=2 Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.398154 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.447060 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-utilities\") pod \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.447154 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjp7w\" (UniqueName: \"kubernetes.io/projected/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-kube-api-access-gjp7w\") pod \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.447216 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-catalog-content\") pod \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\" (UID: \"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af\") " Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.448185 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-utilities" (OuterVolumeSpecName: "utilities") pod "38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" (UID: "38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.457937 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-kube-api-access-gjp7w" (OuterVolumeSpecName: "kube-api-access-gjp7w") pod "38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" (UID: "38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af"). InnerVolumeSpecName "kube-api-access-gjp7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.549620 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.550016 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjp7w\" (UniqueName: \"kubernetes.io/projected/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-kube-api-access-gjp7w\") on node \"crc\" DevicePath \"\"" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.577241 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" (UID: "38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.652004 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.941703 4675 generic.go:334] "Generic (PLEG): container finished" podID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerID="8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad" exitCode=0 Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.941808 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdcj" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.941802 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdcj" event={"ID":"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af","Type":"ContainerDied","Data":"8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad"} Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.942872 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdcj" event={"ID":"38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af","Type":"ContainerDied","Data":"c8776503ac16f3b1282ea3e2961476343541cdc042151395171a1dc694350c3e"} Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.942897 4675 scope.go:117] "RemoveContainer" containerID="8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.978766 4675 scope.go:117] "RemoveContainer" containerID="e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2" Feb 19 09:10:17 crc kubenswrapper[4675]: I0219 09:10:17.993408 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frdcj"] Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.006694 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frdcj"] Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.009143 4675 scope.go:117] "RemoveContainer" containerID="1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762" Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.049963 4675 scope.go:117] "RemoveContainer" containerID="8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad" Feb 19 09:10:18 crc kubenswrapper[4675]: E0219 09:10:18.050694 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad\": container with ID starting with 8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad not found: ID does not exist" containerID="8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad" Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.050733 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad"} err="failed to get container status \"8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad\": rpc error: code = NotFound desc = could not find container \"8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad\": container with ID starting with 8145c86dcdcd322792f55b9980e8d45eecd82a5001e14d95f34f23d1bb0630ad not found: ID does not exist" Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.050756 4675 scope.go:117] "RemoveContainer" containerID="e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2" Feb 19 09:10:18 crc kubenswrapper[4675]: E0219 09:10:18.051216 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2\": container with ID starting with e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2 not found: ID does not exist" containerID="e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2" Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.051296 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2"} err="failed to get container status \"e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2\": rpc error: code = NotFound desc = could not find container \"e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2\": container with ID starting with e7e705be758c350a380563576808d7c7a64423437c1d70d28ed15e7b672e49e2 not found: ID does not exist" Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.051333 4675 scope.go:117] "RemoveContainer" containerID="1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762" Feb 19 09:10:18 crc kubenswrapper[4675]: E0219 09:10:18.052169 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762\": container with ID starting with 1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762 not found: ID does not exist" containerID="1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762" Feb 19 09:10:18 crc kubenswrapper[4675]: I0219 09:10:18.052248 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762"} err="failed to get container status \"1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762\": rpc error: code = NotFound desc = could not find container \"1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762\": container with ID starting with 1a9ac278efa9b5874294d770446b237809804563f84a0a1c9f43b03b21326762 not found: ID does not exist" Feb 19 09:10:19 crc kubenswrapper[4675]: I0219 09:10:19.114729 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" path="/var/lib/kubelet/pods/38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af/volumes" Feb 19 09:10:40 crc kubenswrapper[4675]: I0219 09:10:40.040026 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-hl8ff"] Feb 19 09:10:40 crc kubenswrapper[4675]: I0219 09:10:40.052002 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-hl8ff"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.039063 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-00eb-account-create-update-h6l8c"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.048793 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-l5kvq"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.060305 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gtrrk"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.070810 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-c09a-account-create-update-dx5kw"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.080471 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3afe-account-create-update-ptbwn"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.088712 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-00eb-account-create-update-h6l8c"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.097145 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3afe-account-create-update-ptbwn"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.113929 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20fa7039-8505-420a-a55f-0214b8c5fb52" path="/var/lib/kubelet/pods/20fa7039-8505-420a-a55f-0214b8c5fb52/volumes" Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.115202 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51845ac1-8df1-4f01-860e-c0a0d0eb0508" path="/var/lib/kubelet/pods/51845ac1-8df1-4f01-860e-c0a0d0eb0508/volumes" Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.115971 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99cb40fe-72ba-4abd-bbdb-f21f7f37af3f" path="/var/lib/kubelet/pods/99cb40fe-72ba-4abd-bbdb-f21f7f37af3f/volumes" Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.116576 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gtrrk"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.116617 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-c09a-account-create-update-dx5kw"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.121306 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-l5kvq"] Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.744045 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:10:41 crc kubenswrapper[4675]: I0219 09:10:41.744099 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:10:43 crc kubenswrapper[4675]: I0219 09:10:43.116167 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d990782-ad62-4a59-8f00-02917c2c957c" path="/var/lib/kubelet/pods/1d990782-ad62-4a59-8f00-02917c2c957c/volumes" Feb 19 09:10:43 crc kubenswrapper[4675]: I0219 09:10:43.117250 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de7aed7-aa20-4ba7-95dc-4d96b45c0caa" path="/var/lib/kubelet/pods/2de7aed7-aa20-4ba7-95dc-4d96b45c0caa/volumes" Feb 19 09:10:43 crc kubenswrapper[4675]: I0219 09:10:43.118193 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b749ff9b-32ac-484b-a722-c42543ebc3b9" path="/var/lib/kubelet/pods/b749ff9b-32ac-484b-a722-c42543ebc3b9/volumes" Feb 19 09:10:52 crc kubenswrapper[4675]: I0219 09:10:52.341143 4675 scope.go:117] "RemoveContainer" containerID="0e2a9adc7b5f959b07520dcde45292ed36759b337879b77300141ab7dd29ac13" Feb 19 09:10:52 crc kubenswrapper[4675]: I0219 09:10:52.382080 4675 scope.go:117] "RemoveContainer" containerID="f82a995bae45cad532588f3de77fe1de7b607c5b389a6b0069a7aa4893b3a535" Feb 19 09:10:52 crc kubenswrapper[4675]: I0219 09:10:52.419924 4675 scope.go:117] "RemoveContainer" containerID="b4a058cb8fa7120b3594062632f768f4f2e996b0c7cf5bcfd8129a15299b3e48" Feb 19 09:10:52 crc kubenswrapper[4675]: I0219 09:10:52.463694 4675 scope.go:117] "RemoveContainer" containerID="1a2ce06e34a659ff629585bf0c6c451d0c806958da9acd153dd3590c202efcff" Feb 19 09:10:52 crc kubenswrapper[4675]: I0219 09:10:52.529656 4675 scope.go:117] "RemoveContainer" containerID="5a8c48c07e166c1ef8a78d91ad8dbf9f60a54c6aaa936483addda5d01b11c391" Feb 19 09:10:52 crc kubenswrapper[4675]: I0219 09:10:52.559388 4675 scope.go:117] "RemoveContainer" containerID="25989343d3a8a6300cf62f853a3f65f9a5f665095adfe2e3cc15f2e25b488b8f" Feb 19 09:11:05 crc kubenswrapper[4675]: I0219 09:11:05.381360 4675 generic.go:334] "Generic (PLEG): container finished" podID="31e7b53a-04a2-4fd3-b18e-893da088a60f" containerID="d1b13e2c2b5ce477faa4320a9eb9b3f0045c49fab4802c22767a1d29a0f5138e" exitCode=0 Feb 19 09:11:05 crc kubenswrapper[4675]: I0219 09:11:05.381449 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" event={"ID":"31e7b53a-04a2-4fd3-b18e-893da088a60f","Type":"ContainerDied","Data":"d1b13e2c2b5ce477faa4320a9eb9b3f0045c49fab4802c22767a1d29a0f5138e"} Feb 19 09:11:06 crc kubenswrapper[4675]: I0219 09:11:06.930031 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.039374 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xgg6\" (UniqueName: \"kubernetes.io/projected/31e7b53a-04a2-4fd3-b18e-893da088a60f-kube-api-access-7xgg6\") pod \"31e7b53a-04a2-4fd3-b18e-893da088a60f\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.039821 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-ssh-key-openstack-edpm-ipam\") pod \"31e7b53a-04a2-4fd3-b18e-893da088a60f\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.039993 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-inventory\") pod \"31e7b53a-04a2-4fd3-b18e-893da088a60f\" (UID: \"31e7b53a-04a2-4fd3-b18e-893da088a60f\") " Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.045511 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-4bphc"] Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.055084 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e7b53a-04a2-4fd3-b18e-893da088a60f-kube-api-access-7xgg6" (OuterVolumeSpecName: "kube-api-access-7xgg6") pod "31e7b53a-04a2-4fd3-b18e-893da088a60f" (UID: "31e7b53a-04a2-4fd3-b18e-893da088a60f"). InnerVolumeSpecName "kube-api-access-7xgg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.062337 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-4bphc"] Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.075797 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-inventory" (OuterVolumeSpecName: "inventory") pod "31e7b53a-04a2-4fd3-b18e-893da088a60f" (UID: "31e7b53a-04a2-4fd3-b18e-893da088a60f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.081593 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "31e7b53a-04a2-4fd3-b18e-893da088a60f" (UID: "31e7b53a-04a2-4fd3-b18e-893da088a60f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.115441 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd86c751-bf59-47e4-81c4-dfaeb157fa76" path="/var/lib/kubelet/pods/fd86c751-bf59-47e4-81c4-dfaeb157fa76/volumes" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.142313 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xgg6\" (UniqueName: \"kubernetes.io/projected/31e7b53a-04a2-4fd3-b18e-893da088a60f-kube-api-access-7xgg6\") on node \"crc\" DevicePath \"\"" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.142353 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.142367 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e7b53a-04a2-4fd3-b18e-893da088a60f-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.401736 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" event={"ID":"31e7b53a-04a2-4fd3-b18e-893da088a60f","Type":"ContainerDied","Data":"0eb6a3c9b9561341f6c99696e74c7ef6ec7a7143d90d8ce6d09bdb51487bf34e"} Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.401783 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eb6a3c9b9561341f6c99696e74c7ef6ec7a7143d90d8ce6d09bdb51487bf34e" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.401840 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fr44t" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.493267 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k"] Feb 19 09:11:07 crc kubenswrapper[4675]: E0219 09:11:07.493821 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="registry-server" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.493843 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="registry-server" Feb 19 09:11:07 crc kubenswrapper[4675]: E0219 09:11:07.493861 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="extract-utilities" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.493869 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="extract-utilities" Feb 19 09:11:07 crc kubenswrapper[4675]: E0219 09:11:07.493887 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="extract-content" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.493896 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="extract-content" Feb 19 09:11:07 crc kubenswrapper[4675]: E0219 09:11:07.493929 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e7b53a-04a2-4fd3-b18e-893da088a60f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.493938 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e7b53a-04a2-4fd3-b18e-893da088a60f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.494149 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e7b53a-04a2-4fd3-b18e-893da088a60f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.494177 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="38b8e6ef-68a8-4e97-9ad1-f78ee3b9e2af" containerName="registry-server" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.495025 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.499293 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.499323 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.499465 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.501373 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.508302 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k"] Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.549078 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.549237 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.549310 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5rnn\" (UniqueName: \"kubernetes.io/projected/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-kube-api-access-d5rnn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.651371 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5rnn\" (UniqueName: \"kubernetes.io/projected/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-kube-api-access-d5rnn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.651529 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.651677 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.661235 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.664212 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.679493 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5rnn\" (UniqueName: \"kubernetes.io/projected/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-kube-api-access-d5rnn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:07 crc kubenswrapper[4675]: I0219 09:11:07.816337 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:11:08 crc kubenswrapper[4675]: I0219 09:11:08.330067 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k"] Feb 19 09:11:08 crc kubenswrapper[4675]: I0219 09:11:08.339886 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:11:08 crc kubenswrapper[4675]: I0219 09:11:08.411468 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" event={"ID":"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f","Type":"ContainerStarted","Data":"9f4e25118831bec339407c0d047f3a9765163ca4e4bda61e204293a8bb70eaf5"} Feb 19 09:11:09 crc kubenswrapper[4675]: I0219 09:11:09.422248 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" event={"ID":"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f","Type":"ContainerStarted","Data":"d8d91172b638b8c79c2ad9ee53e4146ed1600ec9921242af591bc45b88d45cf8"} Feb 19 09:11:09 crc kubenswrapper[4675]: I0219 09:11:09.441452 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" podStartSLOduration=1.9909196759999999 podStartE2EDuration="2.441431231s" podCreationTimestamp="2026-02-19 09:11:07 +0000 UTC" firstStartedPulling="2026-02-19 09:11:08.339601901 +0000 UTC m=+1649.966692169" lastFinishedPulling="2026-02-19 09:11:08.790113426 +0000 UTC m=+1650.417203724" observedRunningTime="2026-02-19 09:11:09.436861172 +0000 UTC m=+1651.063951440" watchObservedRunningTime="2026-02-19 09:11:09.441431231 +0000 UTC m=+1651.068521499" Feb 19 09:11:10 crc kubenswrapper[4675]: I0219 09:11:10.036455 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1078-account-create-update-d8w87"] Feb 19 09:11:10 crc kubenswrapper[4675]: I0219 09:11:10.047793 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-8hnwf"] Feb 19 09:11:10 crc kubenswrapper[4675]: I0219 09:11:10.057682 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-8hnwf"] Feb 19 09:11:10 crc kubenswrapper[4675]: I0219 09:11:10.067426 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1078-account-create-update-d8w87"] Feb 19 09:11:11 crc kubenswrapper[4675]: I0219 09:11:11.115331 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a454b94-9aec-4a5a-b92c-6039497d605e" path="/var/lib/kubelet/pods/0a454b94-9aec-4a5a-b92c-6039497d605e/volumes" Feb 19 09:11:11 crc kubenswrapper[4675]: I0219 09:11:11.116394 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8f55242-ce77-4241-ac85-e27d12376939" path="/var/lib/kubelet/pods/c8f55242-ce77-4241-ac85-e27d12376939/volumes" Feb 19 09:11:11 crc kubenswrapper[4675]: I0219 09:11:11.743597 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:11:11 crc kubenswrapper[4675]: I0219 09:11:11.743694 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.036580 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-bec0-account-create-update-mdjhn"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.045438 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vhgrp"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.054580 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-06ee-account-create-update-zpdvd"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.064998 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-r8mnm"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.074096 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-06ee-account-create-update-zpdvd"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.082557 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vhgrp"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.090418 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-r8mnm"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.097999 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-bec0-account-create-update-mdjhn"] Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.116011 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03b37867-7bec-498a-8ef7-2de4403950e8" path="/var/lib/kubelet/pods/03b37867-7bec-498a-8ef7-2de4403950e8/volumes" Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.116873 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13feadc4-3e1d-495b-8177-11af75487059" path="/var/lib/kubelet/pods/13feadc4-3e1d-495b-8177-11af75487059/volumes" Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.117866 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37e4b782-e0a5-4b72-8707-2c361c510f90" path="/var/lib/kubelet/pods/37e4b782-e0a5-4b72-8707-2c361c510f90/volumes" Feb 19 09:11:15 crc kubenswrapper[4675]: I0219 09:11:15.118988 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a907b5a0-9709-4bc4-8b3a-7a364bcf66a9" path="/var/lib/kubelet/pods/a907b5a0-9709-4bc4-8b3a-7a364bcf66a9/volumes" Feb 19 09:11:19 crc kubenswrapper[4675]: I0219 09:11:19.032841 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-flrm8"] Feb 19 09:11:19 crc kubenswrapper[4675]: I0219 09:11:19.050157 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-flrm8"] Feb 19 09:11:19 crc kubenswrapper[4675]: I0219 09:11:19.115585 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16be0916-9aea-444d-bc47-6bd7f646cb26" path="/var/lib/kubelet/pods/16be0916-9aea-444d-bc47-6bd7f646cb26/volumes" Feb 19 09:11:41 crc kubenswrapper[4675]: I0219 09:11:41.743492 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:11:41 crc kubenswrapper[4675]: I0219 09:11:41.744314 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:11:41 crc kubenswrapper[4675]: I0219 09:11:41.744369 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:11:41 crc kubenswrapper[4675]: I0219 09:11:41.745170 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:11:41 crc kubenswrapper[4675]: I0219 09:11:41.745219 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" gracePeriod=600 Feb 19 09:11:41 crc kubenswrapper[4675]: E0219 09:11:41.881142 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:11:42 crc kubenswrapper[4675]: I0219 09:11:42.726310 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" exitCode=0 Feb 19 09:11:42 crc kubenswrapper[4675]: I0219 09:11:42.726357 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44"} Feb 19 09:11:42 crc kubenswrapper[4675]: I0219 09:11:42.726396 4675 scope.go:117] "RemoveContainer" containerID="d8cec9b5ad0c194b9e5ae1c687b8aae050e8cf77a0f404f058bb11195c4e24e3" Feb 19 09:11:42 crc kubenswrapper[4675]: I0219 09:11:42.727051 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:11:42 crc kubenswrapper[4675]: E0219 09:11:42.727403 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:11:50 crc kubenswrapper[4675]: I0219 09:11:50.056991 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-g8ngb"] Feb 19 09:11:50 crc kubenswrapper[4675]: I0219 09:11:50.071738 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-g8ngb"] Feb 19 09:11:51 crc kubenswrapper[4675]: I0219 09:11:51.115871 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ee47cd5-e516-4da9-ac5f-40cbedfe103b" path="/var/lib/kubelet/pods/3ee47cd5-e516-4da9-ac5f-40cbedfe103b/volumes" Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.035122 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-wmmx8"] Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.045039 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-wmmx8"] Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.733407 4675 scope.go:117] "RemoveContainer" containerID="b7a9391a34920b4b0e42d0a5d9ffca8cd522ff8098ce387c13da90f27dacb4f5" Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.764554 4675 scope.go:117] "RemoveContainer" containerID="a6ae9629c316ccf48c4bd31750bf148da87a1a6b9a8d434ef56a12e6a6d8652c" Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.803672 4675 scope.go:117] "RemoveContainer" containerID="97a09a59c6405338e32577c28c1578002a771bb6b35b58cf1fc40eb301c9a94d" Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.853109 4675 scope.go:117] "RemoveContainer" containerID="4c2dfb0bcbabe81d458551f6acc664fb9e6bf529f3c18d617d1ced73c0d8670a" Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.917036 4675 scope.go:117] "RemoveContainer" containerID="46466b799428bac83beddb04de2ca14b4c4ef45edfd5734b60933526dfd13ae0" Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.944031 4675 scope.go:117] "RemoveContainer" containerID="d5f80da71accb19bbb513b98cde1cd78644946a01c371aeaf2f124e3ed6dc557" Feb 19 09:11:52 crc kubenswrapper[4675]: I0219 09:11:52.986542 4675 scope.go:117] "RemoveContainer" containerID="4adba8d72c0f9e8863bb604064b1ee3146319595fb9e1e8b88c3bd5825a45d9e" Feb 19 09:11:53 crc kubenswrapper[4675]: I0219 09:11:53.006374 4675 scope.go:117] "RemoveContainer" containerID="4d7565a8ca901db4e12e439fdf1c78814cc7a9eba7b060a0b60df888676b8d14" Feb 19 09:11:53 crc kubenswrapper[4675]: I0219 09:11:53.045006 4675 scope.go:117] "RemoveContainer" containerID="be26d0805a20aa60bddde3bd34f2760db312e2e2b17b5ad4482b4eb5257af821" Feb 19 09:11:53 crc kubenswrapper[4675]: I0219 09:11:53.107116 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:11:53 crc kubenswrapper[4675]: E0219 09:11:53.107560 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:11:53 crc kubenswrapper[4675]: I0219 09:11:53.118277 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0648a197-2cca-4ce8-bc6e-7162fd388de8" path="/var/lib/kubelet/pods/0648a197-2cca-4ce8-bc6e-7162fd388de8/volumes" Feb 19 09:12:00 crc kubenswrapper[4675]: I0219 09:12:00.044888 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-c47wh"] Feb 19 09:12:00 crc kubenswrapper[4675]: I0219 09:12:00.054933 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-c47wh"] Feb 19 09:12:01 crc kubenswrapper[4675]: I0219 09:12:01.116201 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e02b31ff-8f9a-4d65-b6fe-849630c89ede" path="/var/lib/kubelet/pods/e02b31ff-8f9a-4d65-b6fe-849630c89ede/volumes" Feb 19 09:12:02 crc kubenswrapper[4675]: I0219 09:12:02.031603 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-wz69f"] Feb 19 09:12:02 crc kubenswrapper[4675]: I0219 09:12:02.040753 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-wz69f"] Feb 19 09:12:03 crc kubenswrapper[4675]: I0219 09:12:03.115675 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824b7d51-57cb-459f-8876-9d88025ff60f" path="/var/lib/kubelet/pods/824b7d51-57cb-459f-8876-9d88025ff60f/volumes" Feb 19 09:12:07 crc kubenswrapper[4675]: I0219 09:12:07.103231 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:12:07 crc kubenswrapper[4675]: E0219 09:12:07.104263 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:12:14 crc kubenswrapper[4675]: I0219 09:12:14.031794 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-cbncs"] Feb 19 09:12:14 crc kubenswrapper[4675]: I0219 09:12:14.043923 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-cbncs"] Feb 19 09:12:14 crc kubenswrapper[4675]: I0219 09:12:14.542463 4675 generic.go:334] "Generic (PLEG): container finished" podID="8fa31ca4-5356-4a67-ac53-6bc4d9f0780f" containerID="d8d91172b638b8c79c2ad9ee53e4146ed1600ec9921242af591bc45b88d45cf8" exitCode=0 Feb 19 09:12:14 crc kubenswrapper[4675]: I0219 09:12:14.542555 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" event={"ID":"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f","Type":"ContainerDied","Data":"d8d91172b638b8c79c2ad9ee53e4146ed1600ec9921242af591bc45b88d45cf8"} Feb 19 09:12:15 crc kubenswrapper[4675]: I0219 09:12:15.114863 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da9f4b70-4c48-4658-b2ae-b2d91b58f23c" path="/var/lib/kubelet/pods/da9f4b70-4c48-4658-b2ae-b2d91b58f23c/volumes" Feb 19 09:12:15 crc kubenswrapper[4675]: I0219 09:12:15.953605 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.078948 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-ssh-key-openstack-edpm-ipam\") pod \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.079521 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-inventory\") pod \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.079561 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5rnn\" (UniqueName: \"kubernetes.io/projected/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-kube-api-access-d5rnn\") pod \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\" (UID: \"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f\") " Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.087613 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-kube-api-access-d5rnn" (OuterVolumeSpecName: "kube-api-access-d5rnn") pod "8fa31ca4-5356-4a67-ac53-6bc4d9f0780f" (UID: "8fa31ca4-5356-4a67-ac53-6bc4d9f0780f"). InnerVolumeSpecName "kube-api-access-d5rnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.110071 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-inventory" (OuterVolumeSpecName: "inventory") pod "8fa31ca4-5356-4a67-ac53-6bc4d9f0780f" (UID: "8fa31ca4-5356-4a67-ac53-6bc4d9f0780f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.111250 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8fa31ca4-5356-4a67-ac53-6bc4d9f0780f" (UID: "8fa31ca4-5356-4a67-ac53-6bc4d9f0780f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.182454 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.182499 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.182509 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5rnn\" (UniqueName: \"kubernetes.io/projected/8fa31ca4-5356-4a67-ac53-6bc4d9f0780f-kube-api-access-d5rnn\") on node \"crc\" DevicePath \"\"" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.560383 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" event={"ID":"8fa31ca4-5356-4a67-ac53-6bc4d9f0780f","Type":"ContainerDied","Data":"9f4e25118831bec339407c0d047f3a9765163ca4e4bda61e204293a8bb70eaf5"} Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.560431 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f4e25118831bec339407c0d047f3a9765163ca4e4bda61e204293a8bb70eaf5" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.560444 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.643536 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj"] Feb 19 09:12:16 crc kubenswrapper[4675]: E0219 09:12:16.644165 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa31ca4-5356-4a67-ac53-6bc4d9f0780f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.644190 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa31ca4-5356-4a67-ac53-6bc4d9f0780f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.644402 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa31ca4-5356-4a67-ac53-6bc4d9f0780f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.645450 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.649675 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.649781 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.649890 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.652853 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.658947 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj"] Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.693002 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.693308 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.693475 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd9v2\" (UniqueName: \"kubernetes.io/projected/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-kube-api-access-cd9v2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.795667 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.795725 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd9v2\" (UniqueName: \"kubernetes.io/projected/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-kube-api-access-cd9v2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.795887 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.810571 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.810593 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.813527 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd9v2\" (UniqueName: \"kubernetes.io/projected/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-kube-api-access-cd9v2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:16 crc kubenswrapper[4675]: I0219 09:12:16.967449 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:17 crc kubenswrapper[4675]: I0219 09:12:17.481480 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj"] Feb 19 09:12:17 crc kubenswrapper[4675]: I0219 09:12:17.570459 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" event={"ID":"af6f5807-47f8-4e48-b8c0-b04eeb4bb160","Type":"ContainerStarted","Data":"e3c6cb1f4c565d6457636da20a8534313337f5dba99620f3ebd472a380fc53d6"} Feb 19 09:12:18 crc kubenswrapper[4675]: I0219 09:12:18.584512 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" event={"ID":"af6f5807-47f8-4e48-b8c0-b04eeb4bb160","Type":"ContainerStarted","Data":"d1ca74d371025bbd1e27fd84cb2fd44c0e6121f80d7938b0ac3e03a8b0cda792"} Feb 19 09:12:18 crc kubenswrapper[4675]: I0219 09:12:18.615314 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" podStartSLOduration=2.227572257 podStartE2EDuration="2.615291177s" podCreationTimestamp="2026-02-19 09:12:16 +0000 UTC" firstStartedPulling="2026-02-19 09:12:17.486932963 +0000 UTC m=+1719.114023231" lastFinishedPulling="2026-02-19 09:12:17.874651883 +0000 UTC m=+1719.501742151" observedRunningTime="2026-02-19 09:12:18.614548758 +0000 UTC m=+1720.241639026" watchObservedRunningTime="2026-02-19 09:12:18.615291177 +0000 UTC m=+1720.242381445" Feb 19 09:12:21 crc kubenswrapper[4675]: I0219 09:12:21.035334 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-wzzsx"] Feb 19 09:12:21 crc kubenswrapper[4675]: I0219 09:12:21.046053 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-wzzsx"] Feb 19 09:12:21 crc kubenswrapper[4675]: I0219 09:12:21.114855 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d78a6299-1816-4500-9e03-a004f52262f8" path="/var/lib/kubelet/pods/d78a6299-1816-4500-9e03-a004f52262f8/volumes" Feb 19 09:12:22 crc kubenswrapper[4675]: I0219 09:12:22.103300 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:12:22 crc kubenswrapper[4675]: E0219 09:12:22.103596 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:12:22 crc kubenswrapper[4675]: I0219 09:12:22.617322 4675 generic.go:334] "Generic (PLEG): container finished" podID="af6f5807-47f8-4e48-b8c0-b04eeb4bb160" containerID="d1ca74d371025bbd1e27fd84cb2fd44c0e6121f80d7938b0ac3e03a8b0cda792" exitCode=0 Feb 19 09:12:22 crc kubenswrapper[4675]: I0219 09:12:22.617367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" event={"ID":"af6f5807-47f8-4e48-b8c0-b04eeb4bb160","Type":"ContainerDied","Data":"d1ca74d371025bbd1e27fd84cb2fd44c0e6121f80d7938b0ac3e03a8b0cda792"} Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.173260 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.351612 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-inventory\") pod \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.352187 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-ssh-key-openstack-edpm-ipam\") pod \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.352301 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd9v2\" (UniqueName: \"kubernetes.io/projected/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-kube-api-access-cd9v2\") pod \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\" (UID: \"af6f5807-47f8-4e48-b8c0-b04eeb4bb160\") " Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.357908 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-kube-api-access-cd9v2" (OuterVolumeSpecName: "kube-api-access-cd9v2") pod "af6f5807-47f8-4e48-b8c0-b04eeb4bb160" (UID: "af6f5807-47f8-4e48-b8c0-b04eeb4bb160"). InnerVolumeSpecName "kube-api-access-cd9v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.383220 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "af6f5807-47f8-4e48-b8c0-b04eeb4bb160" (UID: "af6f5807-47f8-4e48-b8c0-b04eeb4bb160"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.389324 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-inventory" (OuterVolumeSpecName: "inventory") pod "af6f5807-47f8-4e48-b8c0-b04eeb4bb160" (UID: "af6f5807-47f8-4e48-b8c0-b04eeb4bb160"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.454486 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.454527 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd9v2\" (UniqueName: \"kubernetes.io/projected/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-kube-api-access-cd9v2\") on node \"crc\" DevicePath \"\"" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.454540 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af6f5807-47f8-4e48-b8c0-b04eeb4bb160-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.640852 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" event={"ID":"af6f5807-47f8-4e48-b8c0-b04eeb4bb160","Type":"ContainerDied","Data":"e3c6cb1f4c565d6457636da20a8534313337f5dba99620f3ebd472a380fc53d6"} Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.640899 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3c6cb1f4c565d6457636da20a8534313337f5dba99620f3ebd472a380fc53d6" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.640909 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.724448 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7"] Feb 19 09:12:24 crc kubenswrapper[4675]: E0219 09:12:24.725038 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6f5807-47f8-4e48-b8c0-b04eeb4bb160" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.725061 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6f5807-47f8-4e48-b8c0-b04eeb4bb160" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.725364 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="af6f5807-47f8-4e48-b8c0-b04eeb4bb160" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.726304 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.749575 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7"] Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.756653 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.756859 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.757123 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.758031 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.862128 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.862525 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-552n6\" (UniqueName: \"kubernetes.io/projected/55d678f4-667d-4d16-8f40-4faeab6d6c1a-kube-api-access-552n6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.862561 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.964401 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.964469 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-552n6\" (UniqueName: \"kubernetes.io/projected/55d678f4-667d-4d16-8f40-4faeab6d6c1a-kube-api-access-552n6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.964508 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.968567 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.968715 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:24 crc kubenswrapper[4675]: I0219 09:12:24.985566 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-552n6\" (UniqueName: \"kubernetes.io/projected/55d678f4-667d-4d16-8f40-4faeab6d6c1a-kube-api-access-552n6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g5kn7\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:25 crc kubenswrapper[4675]: I0219 09:12:25.075259 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:12:25 crc kubenswrapper[4675]: I0219 09:12:25.587561 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7"] Feb 19 09:12:25 crc kubenswrapper[4675]: I0219 09:12:25.650372 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" event={"ID":"55d678f4-667d-4d16-8f40-4faeab6d6c1a","Type":"ContainerStarted","Data":"7a55791addcbac3b56c9271fb8dde7e13e5ea66f1fbc326c949bb13564647618"} Feb 19 09:12:26 crc kubenswrapper[4675]: I0219 09:12:26.661356 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" event={"ID":"55d678f4-667d-4d16-8f40-4faeab6d6c1a","Type":"ContainerStarted","Data":"8244b3d1fc7d004283ae4356e633bd946ac8950fa644bf31ff1fd317544709d8"} Feb 19 09:12:26 crc kubenswrapper[4675]: I0219 09:12:26.682361 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" podStartSLOduration=2.080109413 podStartE2EDuration="2.682340483s" podCreationTimestamp="2026-02-19 09:12:24 +0000 UTC" firstStartedPulling="2026-02-19 09:12:25.599385947 +0000 UTC m=+1727.226476215" lastFinishedPulling="2026-02-19 09:12:26.201617017 +0000 UTC m=+1727.828707285" observedRunningTime="2026-02-19 09:12:26.675880231 +0000 UTC m=+1728.302970499" watchObservedRunningTime="2026-02-19 09:12:26.682340483 +0000 UTC m=+1728.309430751" Feb 19 09:12:33 crc kubenswrapper[4675]: I0219 09:12:33.103285 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:12:33 crc kubenswrapper[4675]: E0219 09:12:33.103921 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:12:44 crc kubenswrapper[4675]: I0219 09:12:44.103272 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:12:44 crc kubenswrapper[4675]: E0219 09:12:44.104393 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:12:53 crc kubenswrapper[4675]: I0219 09:12:53.212315 4675 scope.go:117] "RemoveContainer" containerID="0a37609f47684f1831ddd07f069c7cf31f5b88a09a2bffbde4f3f4c461f9dca1" Feb 19 09:12:53 crc kubenswrapper[4675]: I0219 09:12:53.257865 4675 scope.go:117] "RemoveContainer" containerID="9cf858bc93f6aefa286d61064d358262d8ef1f55263340fc054c72ac2ca4183e" Feb 19 09:12:53 crc kubenswrapper[4675]: I0219 09:12:53.287025 4675 scope.go:117] "RemoveContainer" containerID="0213029a75f69fc51b42589a3d914c1d1fdc5aab2b75fed314ed84ac198bedaf" Feb 19 09:12:53 crc kubenswrapper[4675]: I0219 09:12:53.359899 4675 scope.go:117] "RemoveContainer" containerID="1ce20a26d60177b6b285b56136a8da2029d771e37237d9234f57d0e3642d0c9f" Feb 19 09:12:53 crc kubenswrapper[4675]: I0219 09:12:53.426959 4675 scope.go:117] "RemoveContainer" containerID="7f7a03e416d35621ac5cca2cbccaa49a47d259749f7a9fb584360690688e6bb2" Feb 19 09:12:56 crc kubenswrapper[4675]: I0219 09:12:56.105010 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:12:56 crc kubenswrapper[4675]: E0219 09:12:56.105958 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.057747 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-q2tv7"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.069691 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0a16-account-create-update-n7pz4"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.083074 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-e4fa-account-create-update-rbg2p"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.092278 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-crs9q"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.100303 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vk5pq"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.113689 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-q2tv7"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.127757 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-crs9q"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.131541 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-e4fa-account-create-update-rbg2p"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.141802 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vk5pq"] Feb 19 09:12:57 crc kubenswrapper[4675]: I0219 09:12:57.148284 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0a16-account-create-update-n7pz4"] Feb 19 09:12:58 crc kubenswrapper[4675]: I0219 09:12:58.030064 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-61a4-account-create-update-rj4bg"] Feb 19 09:12:58 crc kubenswrapper[4675]: I0219 09:12:58.039175 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-61a4-account-create-update-rj4bg"] Feb 19 09:12:59 crc kubenswrapper[4675]: I0219 09:12:59.147794 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="179577e4-e654-46ef-b92e-47891c2ff281" path="/var/lib/kubelet/pods/179577e4-e654-46ef-b92e-47891c2ff281/volumes" Feb 19 09:12:59 crc kubenswrapper[4675]: I0219 09:12:59.149327 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32253b6f-fc03-402c-bd55-8d49e4e3a7f1" path="/var/lib/kubelet/pods/32253b6f-fc03-402c-bd55-8d49e4e3a7f1/volumes" Feb 19 09:12:59 crc kubenswrapper[4675]: I0219 09:12:59.150010 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="461c89f2-a273-4964-b6fa-e6a858c7b97c" path="/var/lib/kubelet/pods/461c89f2-a273-4964-b6fa-e6a858c7b97c/volumes" Feb 19 09:12:59 crc kubenswrapper[4675]: I0219 09:12:59.150687 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57dedd62-6c43-4bc5-a2f8-e98d20a8e73e" path="/var/lib/kubelet/pods/57dedd62-6c43-4bc5-a2f8-e98d20a8e73e/volumes" Feb 19 09:12:59 crc kubenswrapper[4675]: I0219 09:12:59.151902 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b190a6d-42fd-487a-a90d-423a63e23603" path="/var/lib/kubelet/pods/8b190a6d-42fd-487a-a90d-423a63e23603/volumes" Feb 19 09:12:59 crc kubenswrapper[4675]: I0219 09:12:59.152592 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9a8c34-ceea-4aac-91ca-ff7a8064c73b" path="/var/lib/kubelet/pods/de9a8c34-ceea-4aac-91ca-ff7a8064c73b/volumes" Feb 19 09:13:00 crc kubenswrapper[4675]: I0219 09:13:00.958278 4675 generic.go:334] "Generic (PLEG): container finished" podID="55d678f4-667d-4d16-8f40-4faeab6d6c1a" containerID="8244b3d1fc7d004283ae4356e633bd946ac8950fa644bf31ff1fd317544709d8" exitCode=0 Feb 19 09:13:00 crc kubenswrapper[4675]: I0219 09:13:00.958345 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" event={"ID":"55d678f4-667d-4d16-8f40-4faeab6d6c1a","Type":"ContainerDied","Data":"8244b3d1fc7d004283ae4356e633bd946ac8950fa644bf31ff1fd317544709d8"} Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.376525 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.514260 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-inventory\") pod \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.514329 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-552n6\" (UniqueName: \"kubernetes.io/projected/55d678f4-667d-4d16-8f40-4faeab6d6c1a-kube-api-access-552n6\") pod \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.515011 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-ssh-key-openstack-edpm-ipam\") pod \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\" (UID: \"55d678f4-667d-4d16-8f40-4faeab6d6c1a\") " Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.524843 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55d678f4-667d-4d16-8f40-4faeab6d6c1a-kube-api-access-552n6" (OuterVolumeSpecName: "kube-api-access-552n6") pod "55d678f4-667d-4d16-8f40-4faeab6d6c1a" (UID: "55d678f4-667d-4d16-8f40-4faeab6d6c1a"). InnerVolumeSpecName "kube-api-access-552n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.545475 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "55d678f4-667d-4d16-8f40-4faeab6d6c1a" (UID: "55d678f4-667d-4d16-8f40-4faeab6d6c1a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.546952 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-inventory" (OuterVolumeSpecName: "inventory") pod "55d678f4-667d-4d16-8f40-4faeab6d6c1a" (UID: "55d678f4-667d-4d16-8f40-4faeab6d6c1a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.617244 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.617298 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55d678f4-667d-4d16-8f40-4faeab6d6c1a-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.617308 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-552n6\" (UniqueName: \"kubernetes.io/projected/55d678f4-667d-4d16-8f40-4faeab6d6c1a-kube-api-access-552n6\") on node \"crc\" DevicePath \"\"" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.979739 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" event={"ID":"55d678f4-667d-4d16-8f40-4faeab6d6c1a","Type":"ContainerDied","Data":"7a55791addcbac3b56c9271fb8dde7e13e5ea66f1fbc326c949bb13564647618"} Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.979792 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g5kn7" Feb 19 09:13:02 crc kubenswrapper[4675]: I0219 09:13:02.979797 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a55791addcbac3b56c9271fb8dde7e13e5ea66f1fbc326c949bb13564647618" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.130067 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z"] Feb 19 09:13:03 crc kubenswrapper[4675]: E0219 09:13:03.132060 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d678f4-667d-4d16-8f40-4faeab6d6c1a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.132096 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d678f4-667d-4d16-8f40-4faeab6d6c1a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.132343 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="55d678f4-667d-4d16-8f40-4faeab6d6c1a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.133241 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.135447 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.135591 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.135899 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.136023 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.150411 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z"] Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.330072 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.330299 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.330434 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4qt9\" (UniqueName: \"kubernetes.io/projected/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-kube-api-access-l4qt9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.433854 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.434022 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.434066 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4qt9\" (UniqueName: \"kubernetes.io/projected/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-kube-api-access-l4qt9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.443653 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.445285 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.451910 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4qt9\" (UniqueName: \"kubernetes.io/projected/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-kube-api-access-l4qt9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.455917 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:03 crc kubenswrapper[4675]: I0219 09:13:03.967506 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z"] Feb 19 09:13:04 crc kubenswrapper[4675]: I0219 09:13:04.996799 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" event={"ID":"380f6f00-e968-4289-8ea9-ed1c8ef7ff59","Type":"ContainerStarted","Data":"175529d8c864b976833474defc270d12ba3b2f564388566a6933a91050be9a4e"} Feb 19 09:13:04 crc kubenswrapper[4675]: I0219 09:13:04.997290 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" event={"ID":"380f6f00-e968-4289-8ea9-ed1c8ef7ff59","Type":"ContainerStarted","Data":"8b77791a963b6da536edc545da9bca2099d5694c80557d2b743c7318bc152d69"} Feb 19 09:13:05 crc kubenswrapper[4675]: I0219 09:13:05.017461 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" podStartSLOduration=1.588799429 podStartE2EDuration="2.017442738s" podCreationTimestamp="2026-02-19 09:13:03 +0000 UTC" firstStartedPulling="2026-02-19 09:13:03.989367353 +0000 UTC m=+1765.616457621" lastFinishedPulling="2026-02-19 09:13:04.418010672 +0000 UTC m=+1766.045100930" observedRunningTime="2026-02-19 09:13:05.010218835 +0000 UTC m=+1766.637309093" watchObservedRunningTime="2026-02-19 09:13:05.017442738 +0000 UTC m=+1766.644533006" Feb 19 09:13:10 crc kubenswrapper[4675]: I0219 09:13:10.104062 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:13:10 crc kubenswrapper[4675]: E0219 09:13:10.105036 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:13:22 crc kubenswrapper[4675]: I0219 09:13:22.103252 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:13:22 crc kubenswrapper[4675]: E0219 09:13:22.104260 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:13:28 crc kubenswrapper[4675]: I0219 09:13:28.046522 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dlxqd"] Feb 19 09:13:28 crc kubenswrapper[4675]: I0219 09:13:28.055914 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dlxqd"] Feb 19 09:13:29 crc kubenswrapper[4675]: I0219 09:13:29.114271 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6c638bb-53d1-4c7a-889b-a506b89b202e" path="/var/lib/kubelet/pods/f6c638bb-53d1-4c7a-889b-a506b89b202e/volumes" Feb 19 09:13:33 crc kubenswrapper[4675]: I0219 09:13:33.103176 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:13:33 crc kubenswrapper[4675]: E0219 09:13:33.104235 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:13:47 crc kubenswrapper[4675]: I0219 09:13:47.166876 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:13:47 crc kubenswrapper[4675]: E0219 09:13:47.168120 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:13:47 crc kubenswrapper[4675]: I0219 09:13:47.397527 4675 generic.go:334] "Generic (PLEG): container finished" podID="380f6f00-e968-4289-8ea9-ed1c8ef7ff59" containerID="175529d8c864b976833474defc270d12ba3b2f564388566a6933a91050be9a4e" exitCode=0 Feb 19 09:13:47 crc kubenswrapper[4675]: I0219 09:13:47.397571 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" event={"ID":"380f6f00-e968-4289-8ea9-ed1c8ef7ff59","Type":"ContainerDied","Data":"175529d8c864b976833474defc270d12ba3b2f564388566a6933a91050be9a4e"} Feb 19 09:13:48 crc kubenswrapper[4675]: I0219 09:13:48.872838 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.015487 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-ssh-key-openstack-edpm-ipam\") pod \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.016712 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4qt9\" (UniqueName: \"kubernetes.io/projected/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-kube-api-access-l4qt9\") pod \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.016998 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-inventory\") pod \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\" (UID: \"380f6f00-e968-4289-8ea9-ed1c8ef7ff59\") " Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.022492 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-kube-api-access-l4qt9" (OuterVolumeSpecName: "kube-api-access-l4qt9") pod "380f6f00-e968-4289-8ea9-ed1c8ef7ff59" (UID: "380f6f00-e968-4289-8ea9-ed1c8ef7ff59"). InnerVolumeSpecName "kube-api-access-l4qt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.042908 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "380f6f00-e968-4289-8ea9-ed1c8ef7ff59" (UID: "380f6f00-e968-4289-8ea9-ed1c8ef7ff59"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.048952 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-inventory" (OuterVolumeSpecName: "inventory") pod "380f6f00-e968-4289-8ea9-ed1c8ef7ff59" (UID: "380f6f00-e968-4289-8ea9-ed1c8ef7ff59"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.119328 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.119378 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4qt9\" (UniqueName: \"kubernetes.io/projected/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-kube-api-access-l4qt9\") on node \"crc\" DevicePath \"\"" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.119390 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/380f6f00-e968-4289-8ea9-ed1c8ef7ff59-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.415151 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" event={"ID":"380f6f00-e968-4289-8ea9-ed1c8ef7ff59","Type":"ContainerDied","Data":"8b77791a963b6da536edc545da9bca2099d5694c80557d2b743c7318bc152d69"} Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.415534 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b77791a963b6da536edc545da9bca2099d5694c80557d2b743c7318bc152d69" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.415240 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.498752 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w7zd2"] Feb 19 09:13:49 crc kubenswrapper[4675]: E0219 09:13:49.499161 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="380f6f00-e968-4289-8ea9-ed1c8ef7ff59" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.499180 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="380f6f00-e968-4289-8ea9-ed1c8ef7ff59" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.499408 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="380f6f00-e968-4289-8ea9-ed1c8ef7ff59" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.500236 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.506209 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.507425 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.509267 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.510103 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.514567 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w7zd2"] Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.628382 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.628711 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjnb2\" (UniqueName: \"kubernetes.io/projected/36e6ad34-b4ec-4e70-895b-37b09f4508aa-kube-api-access-sjnb2\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.628996 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.731899 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjnb2\" (UniqueName: \"kubernetes.io/projected/36e6ad34-b4ec-4e70-895b-37b09f4508aa-kube-api-access-sjnb2\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.732485 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.732542 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.742474 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.742617 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.748285 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjnb2\" (UniqueName: \"kubernetes.io/projected/36e6ad34-b4ec-4e70-895b-37b09f4508aa-kube-api-access-sjnb2\") pod \"ssh-known-hosts-edpm-deployment-w7zd2\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:49 crc kubenswrapper[4675]: I0219 09:13:49.823405 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:13:50 crc kubenswrapper[4675]: I0219 09:13:50.324894 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w7zd2"] Feb 19 09:13:50 crc kubenswrapper[4675]: I0219 09:13:50.422962 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" event={"ID":"36e6ad34-b4ec-4e70-895b-37b09f4508aa","Type":"ContainerStarted","Data":"b499074a294163d732c1a2020cee5da4aeb0bd8c3621daa5b81f0207c48bd884"} Feb 19 09:13:51 crc kubenswrapper[4675]: I0219 09:13:51.042412 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fjxzh"] Feb 19 09:13:51 crc kubenswrapper[4675]: I0219 09:13:51.056133 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fjxzh"] Feb 19 09:13:51 crc kubenswrapper[4675]: I0219 09:13:51.115675 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5077fc1f-d1da-4683-84e8-52afdefcd0e1" path="/var/lib/kubelet/pods/5077fc1f-d1da-4683-84e8-52afdefcd0e1/volumes" Feb 19 09:13:51 crc kubenswrapper[4675]: I0219 09:13:51.433277 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" event={"ID":"36e6ad34-b4ec-4e70-895b-37b09f4508aa","Type":"ContainerStarted","Data":"69d23d79b244becae9d1cdc42eb63f7fac48c7858e4ba69c5b362c90669e6aad"} Feb 19 09:13:51 crc kubenswrapper[4675]: I0219 09:13:51.456288 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" podStartSLOduration=1.904089031 podStartE2EDuration="2.456264223s" podCreationTimestamp="2026-02-19 09:13:49 +0000 UTC" firstStartedPulling="2026-02-19 09:13:50.334269048 +0000 UTC m=+1811.961359306" lastFinishedPulling="2026-02-19 09:13:50.88644423 +0000 UTC m=+1812.513534498" observedRunningTime="2026-02-19 09:13:51.445993495 +0000 UTC m=+1813.073083773" watchObservedRunningTime="2026-02-19 09:13:51.456264223 +0000 UTC m=+1813.083354491" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.563074 4675 scope.go:117] "RemoveContainer" containerID="9a5212464f4459a1bf7820a513a2f4e542f9da83989bf7a03af93f45c8254467" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.587866 4675 scope.go:117] "RemoveContainer" containerID="2bfb00452299892a3c507150c5a2f51c41e62e03a71e43568f7b777267fe6ba0" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.648040 4675 scope.go:117] "RemoveContainer" containerID="ca09d3bbd549d2bc851a6704d9adb6c57182f2f3c86ff524a2f2887c95e77a57" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.693601 4675 scope.go:117] "RemoveContainer" containerID="0a3940db54abe60308aff0e453b5e139c01a362c8b43e4bc11a12997bbfa4822" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.760658 4675 scope.go:117] "RemoveContainer" containerID="e9ff2ea7430ecd53359a998359ec420e76d30023669c91d97a57a414a2354b1b" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.797422 4675 scope.go:117] "RemoveContainer" containerID="ce44f8999c65da848864ac9d9109736b9e5c4a906ed6d0f7c95e31f6f092dc1a" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.838517 4675 scope.go:117] "RemoveContainer" containerID="22c920764004a79203fd0bfb3c5c316f352faa8b894cab5a2e245f6c6ab95eac" Feb 19 09:13:53 crc kubenswrapper[4675]: I0219 09:13:53.861326 4675 scope.go:117] "RemoveContainer" containerID="8405c98fd4eab583741f22c9bcfcf596e2dd54cdbf4d0b8ecca8e43e6115109d" Feb 19 09:13:55 crc kubenswrapper[4675]: I0219 09:13:55.041904 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8lmq6"] Feb 19 09:13:55 crc kubenswrapper[4675]: I0219 09:13:55.049292 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8lmq6"] Feb 19 09:13:55 crc kubenswrapper[4675]: I0219 09:13:55.113441 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102" path="/var/lib/kubelet/pods/3f455bdc-3fe8-4dd4-88e8-a1a24a7f6102/volumes" Feb 19 09:13:58 crc kubenswrapper[4675]: I0219 09:13:58.493013 4675 generic.go:334] "Generic (PLEG): container finished" podID="36e6ad34-b4ec-4e70-895b-37b09f4508aa" containerID="69d23d79b244becae9d1cdc42eb63f7fac48c7858e4ba69c5b362c90669e6aad" exitCode=0 Feb 19 09:13:58 crc kubenswrapper[4675]: I0219 09:13:58.493099 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" event={"ID":"36e6ad34-b4ec-4e70-895b-37b09f4508aa","Type":"ContainerDied","Data":"69d23d79b244becae9d1cdc42eb63f7fac48c7858e4ba69c5b362c90669e6aad"} Feb 19 09:13:59 crc kubenswrapper[4675]: I0219 09:13:59.110216 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:13:59 crc kubenswrapper[4675]: E0219 09:13:59.110558 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:13:59 crc kubenswrapper[4675]: I0219 09:13:59.895423 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.038162 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-ssh-key-openstack-edpm-ipam\") pod \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.038321 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjnb2\" (UniqueName: \"kubernetes.io/projected/36e6ad34-b4ec-4e70-895b-37b09f4508aa-kube-api-access-sjnb2\") pod \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.038486 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-inventory-0\") pod \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\" (UID: \"36e6ad34-b4ec-4e70-895b-37b09f4508aa\") " Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.045310 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36e6ad34-b4ec-4e70-895b-37b09f4508aa-kube-api-access-sjnb2" (OuterVolumeSpecName: "kube-api-access-sjnb2") pod "36e6ad34-b4ec-4e70-895b-37b09f4508aa" (UID: "36e6ad34-b4ec-4e70-895b-37b09f4508aa"). InnerVolumeSpecName "kube-api-access-sjnb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.065473 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "36e6ad34-b4ec-4e70-895b-37b09f4508aa" (UID: "36e6ad34-b4ec-4e70-895b-37b09f4508aa"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.065530 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "36e6ad34-b4ec-4e70-895b-37b09f4508aa" (UID: "36e6ad34-b4ec-4e70-895b-37b09f4508aa"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.140546 4675 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.141029 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36e6ad34-b4ec-4e70-895b-37b09f4508aa-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.141045 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjnb2\" (UniqueName: \"kubernetes.io/projected/36e6ad34-b4ec-4e70-895b-37b09f4508aa-kube-api-access-sjnb2\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.508945 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" event={"ID":"36e6ad34-b4ec-4e70-895b-37b09f4508aa","Type":"ContainerDied","Data":"b499074a294163d732c1a2020cee5da4aeb0bd8c3621daa5b81f0207c48bd884"} Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.508993 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b499074a294163d732c1a2020cee5da4aeb0bd8c3621daa5b81f0207c48bd884" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.509020 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w7zd2" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.590919 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6"] Feb 19 09:14:00 crc kubenswrapper[4675]: E0219 09:14:00.591351 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e6ad34-b4ec-4e70-895b-37b09f4508aa" containerName="ssh-known-hosts-edpm-deployment" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.591376 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e6ad34-b4ec-4e70-895b-37b09f4508aa" containerName="ssh-known-hosts-edpm-deployment" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.591619 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e6ad34-b4ec-4e70-895b-37b09f4508aa" containerName="ssh-known-hosts-edpm-deployment" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.593164 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.596534 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.596561 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.597413 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.598321 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.602890 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6"] Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.753099 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.753172 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.753279 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf5vl\" (UniqueName: \"kubernetes.io/projected/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-kube-api-access-rf5vl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.855057 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.855185 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf5vl\" (UniqueName: \"kubernetes.io/projected/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-kube-api-access-rf5vl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.855251 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.859126 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.859815 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.880782 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf5vl\" (UniqueName: \"kubernetes.io/projected/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-kube-api-access-rf5vl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-692m6\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:00 crc kubenswrapper[4675]: I0219 09:14:00.922323 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:01 crc kubenswrapper[4675]: I0219 09:14:01.426749 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6"] Feb 19 09:14:01 crc kubenswrapper[4675]: I0219 09:14:01.516217 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" event={"ID":"f3bba895-fd3d-44bf-9ea7-e631560bbc0a","Type":"ContainerStarted","Data":"346ee9f01eaaa5b236160f3eabd0e57da707de8b19afcfe75e1ce4d6e170dbd9"} Feb 19 09:14:02 crc kubenswrapper[4675]: I0219 09:14:02.529450 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" event={"ID":"f3bba895-fd3d-44bf-9ea7-e631560bbc0a","Type":"ContainerStarted","Data":"a078d97ed9df8e35d3a97c93572322ab8f9ba60f0e0f9bef1f3e996fa3c9cf2d"} Feb 19 09:14:02 crc kubenswrapper[4675]: I0219 09:14:02.555061 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" podStartSLOduration=2.082627331 podStartE2EDuration="2.555040267s" podCreationTimestamp="2026-02-19 09:14:00 +0000 UTC" firstStartedPulling="2026-02-19 09:14:01.435362561 +0000 UTC m=+1823.062452829" lastFinishedPulling="2026-02-19 09:14:01.907775497 +0000 UTC m=+1823.534865765" observedRunningTime="2026-02-19 09:14:02.548391217 +0000 UTC m=+1824.175481495" watchObservedRunningTime="2026-02-19 09:14:02.555040267 +0000 UTC m=+1824.182130545" Feb 19 09:14:09 crc kubenswrapper[4675]: I0219 09:14:09.591020 4675 generic.go:334] "Generic (PLEG): container finished" podID="f3bba895-fd3d-44bf-9ea7-e631560bbc0a" containerID="a078d97ed9df8e35d3a97c93572322ab8f9ba60f0e0f9bef1f3e996fa3c9cf2d" exitCode=0 Feb 19 09:14:09 crc kubenswrapper[4675]: I0219 09:14:09.591102 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" event={"ID":"f3bba895-fd3d-44bf-9ea7-e631560bbc0a","Type":"ContainerDied","Data":"a078d97ed9df8e35d3a97c93572322ab8f9ba60f0e0f9bef1f3e996fa3c9cf2d"} Feb 19 09:14:10 crc kubenswrapper[4675]: I0219 09:14:10.104178 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:14:10 crc kubenswrapper[4675]: E0219 09:14:10.104789 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:14:10 crc kubenswrapper[4675]: I0219 09:14:10.995258 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.155161 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-inventory\") pod \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.155466 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-ssh-key-openstack-edpm-ipam\") pod \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.155536 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf5vl\" (UniqueName: \"kubernetes.io/projected/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-kube-api-access-rf5vl\") pod \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\" (UID: \"f3bba895-fd3d-44bf-9ea7-e631560bbc0a\") " Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.162033 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-kube-api-access-rf5vl" (OuterVolumeSpecName: "kube-api-access-rf5vl") pod "f3bba895-fd3d-44bf-9ea7-e631560bbc0a" (UID: "f3bba895-fd3d-44bf-9ea7-e631560bbc0a"). InnerVolumeSpecName "kube-api-access-rf5vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.187770 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f3bba895-fd3d-44bf-9ea7-e631560bbc0a" (UID: "f3bba895-fd3d-44bf-9ea7-e631560bbc0a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.200100 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-inventory" (OuterVolumeSpecName: "inventory") pod "f3bba895-fd3d-44bf-9ea7-e631560bbc0a" (UID: "f3bba895-fd3d-44bf-9ea7-e631560bbc0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.258696 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.258734 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.258748 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf5vl\" (UniqueName: \"kubernetes.io/projected/f3bba895-fd3d-44bf-9ea7-e631560bbc0a-kube-api-access-rf5vl\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.610259 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" event={"ID":"f3bba895-fd3d-44bf-9ea7-e631560bbc0a","Type":"ContainerDied","Data":"346ee9f01eaaa5b236160f3eabd0e57da707de8b19afcfe75e1ce4d6e170dbd9"} Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.610314 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="346ee9f01eaaa5b236160f3eabd0e57da707de8b19afcfe75e1ce4d6e170dbd9" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.610333 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-692m6" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.700535 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt"] Feb 19 09:14:11 crc kubenswrapper[4675]: E0219 09:14:11.701039 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3bba895-fd3d-44bf-9ea7-e631560bbc0a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.701061 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3bba895-fd3d-44bf-9ea7-e631560bbc0a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.701274 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3bba895-fd3d-44bf-9ea7-e631560bbc0a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.702135 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.705558 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.705921 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.707080 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.714121 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.717688 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt"] Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.870094 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.870137 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np68g\" (UniqueName: \"kubernetes.io/projected/50eaff9d-656b-4e39-b450-bcc31570bc29-kube-api-access-np68g\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.870182 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.972576 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.972669 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np68g\" (UniqueName: \"kubernetes.io/projected/50eaff9d-656b-4e39-b450-bcc31570bc29-kube-api-access-np68g\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.972725 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.977619 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.979864 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:11 crc kubenswrapper[4675]: I0219 09:14:11.991963 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np68g\" (UniqueName: \"kubernetes.io/projected/50eaff9d-656b-4e39-b450-bcc31570bc29-kube-api-access-np68g\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:12 crc kubenswrapper[4675]: I0219 09:14:12.022538 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:12 crc kubenswrapper[4675]: I0219 09:14:12.345940 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt"] Feb 19 09:14:12 crc kubenswrapper[4675]: I0219 09:14:12.619549 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" event={"ID":"50eaff9d-656b-4e39-b450-bcc31570bc29","Type":"ContainerStarted","Data":"a20b3ce73c176a11654212a7d1c257aa5e8956ca596c81a9c784dc896e7126e4"} Feb 19 09:14:13 crc kubenswrapper[4675]: I0219 09:14:13.629051 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" event={"ID":"50eaff9d-656b-4e39-b450-bcc31570bc29","Type":"ContainerStarted","Data":"82ebad6f2a0ece3d8911b1307b1b2fd4b497b6aefe7a18cb4f28999429e3dbfe"} Feb 19 09:14:13 crc kubenswrapper[4675]: I0219 09:14:13.653774 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" podStartSLOduration=2.117402351 podStartE2EDuration="2.653756609s" podCreationTimestamp="2026-02-19 09:14:11 +0000 UTC" firstStartedPulling="2026-02-19 09:14:12.358927392 +0000 UTC m=+1833.986017670" lastFinishedPulling="2026-02-19 09:14:12.89528166 +0000 UTC m=+1834.522371928" observedRunningTime="2026-02-19 09:14:13.652010972 +0000 UTC m=+1835.279101240" watchObservedRunningTime="2026-02-19 09:14:13.653756609 +0000 UTC m=+1835.280846877" Feb 19 09:14:21 crc kubenswrapper[4675]: I0219 09:14:21.105599 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:14:21 crc kubenswrapper[4675]: E0219 09:14:21.106993 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:14:22 crc kubenswrapper[4675]: I0219 09:14:22.721208 4675 generic.go:334] "Generic (PLEG): container finished" podID="50eaff9d-656b-4e39-b450-bcc31570bc29" containerID="82ebad6f2a0ece3d8911b1307b1b2fd4b497b6aefe7a18cb4f28999429e3dbfe" exitCode=0 Feb 19 09:14:22 crc kubenswrapper[4675]: I0219 09:14:22.721306 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" event={"ID":"50eaff9d-656b-4e39-b450-bcc31570bc29","Type":"ContainerDied","Data":"82ebad6f2a0ece3d8911b1307b1b2fd4b497b6aefe7a18cb4f28999429e3dbfe"} Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.210916 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.337425 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-inventory\") pod \"50eaff9d-656b-4e39-b450-bcc31570bc29\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.337678 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-ssh-key-openstack-edpm-ipam\") pod \"50eaff9d-656b-4e39-b450-bcc31570bc29\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.337754 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np68g\" (UniqueName: \"kubernetes.io/projected/50eaff9d-656b-4e39-b450-bcc31570bc29-kube-api-access-np68g\") pod \"50eaff9d-656b-4e39-b450-bcc31570bc29\" (UID: \"50eaff9d-656b-4e39-b450-bcc31570bc29\") " Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.344353 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50eaff9d-656b-4e39-b450-bcc31570bc29-kube-api-access-np68g" (OuterVolumeSpecName: "kube-api-access-np68g") pod "50eaff9d-656b-4e39-b450-bcc31570bc29" (UID: "50eaff9d-656b-4e39-b450-bcc31570bc29"). InnerVolumeSpecName "kube-api-access-np68g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.368065 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "50eaff9d-656b-4e39-b450-bcc31570bc29" (UID: "50eaff9d-656b-4e39-b450-bcc31570bc29"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.368565 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-inventory" (OuterVolumeSpecName: "inventory") pod "50eaff9d-656b-4e39-b450-bcc31570bc29" (UID: "50eaff9d-656b-4e39-b450-bcc31570bc29"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.439831 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.439877 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np68g\" (UniqueName: \"kubernetes.io/projected/50eaff9d-656b-4e39-b450-bcc31570bc29-kube-api-access-np68g\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.439890 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50eaff9d-656b-4e39-b450-bcc31570bc29-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.742959 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" event={"ID":"50eaff9d-656b-4e39-b450-bcc31570bc29","Type":"ContainerDied","Data":"a20b3ce73c176a11654212a7d1c257aa5e8956ca596c81a9c784dc896e7126e4"} Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.743003 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a20b3ce73c176a11654212a7d1c257aa5e8956ca596c81a9c784dc896e7126e4" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.743062 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.843546 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8"] Feb 19 09:14:24 crc kubenswrapper[4675]: E0219 09:14:24.844239 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50eaff9d-656b-4e39-b450-bcc31570bc29" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.844319 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="50eaff9d-656b-4e39-b450-bcc31570bc29" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.844569 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="50eaff9d-656b-4e39-b450-bcc31570bc29" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.848114 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.851921 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.852386 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.852558 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.852728 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.853050 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.853521 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.853951 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.857203 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.868131 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8"] Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.949921 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950000 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950037 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950096 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950128 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46rb7\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-kube-api-access-46rb7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950229 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950297 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950378 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950432 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950472 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950511 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950627 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950713 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:24 crc kubenswrapper[4675]: I0219 09:14:24.950840 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053108 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053162 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053189 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053225 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053253 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46rb7\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-kube-api-access-46rb7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053290 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053322 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053356 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053391 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053418 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053448 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053480 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053500 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.053540 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.058939 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.059010 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.059265 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.059329 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.059788 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.059908 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.061942 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.062615 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.063691 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.063896 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.064253 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.065273 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.069731 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.071242 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46rb7\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-kube-api-access-46rb7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5cts8\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.166841 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.703000 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8"] Feb 19 09:14:25 crc kubenswrapper[4675]: I0219 09:14:25.753922 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" event={"ID":"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa","Type":"ContainerStarted","Data":"925cacefbb8c679a07b4aa500ea1777b33d07236d2a45b66bcdaa30aa7701b6e"} Feb 19 09:14:26 crc kubenswrapper[4675]: I0219 09:14:26.768285 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" event={"ID":"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa","Type":"ContainerStarted","Data":"69607d8b25f7db5590da483e8d9500965f55034699d379a6ccc28638e7b6f92d"} Feb 19 09:14:26 crc kubenswrapper[4675]: I0219 09:14:26.792039 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" podStartSLOduration=2.404001452 podStartE2EDuration="2.792012244s" podCreationTimestamp="2026-02-19 09:14:24 +0000 UTC" firstStartedPulling="2026-02-19 09:14:25.710625636 +0000 UTC m=+1847.337715904" lastFinishedPulling="2026-02-19 09:14:26.098636428 +0000 UTC m=+1847.725726696" observedRunningTime="2026-02-19 09:14:26.787087241 +0000 UTC m=+1848.414177519" watchObservedRunningTime="2026-02-19 09:14:26.792012244 +0000 UTC m=+1848.419102512" Feb 19 09:14:33 crc kubenswrapper[4675]: I0219 09:14:33.104278 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:14:33 crc kubenswrapper[4675]: E0219 09:14:33.105249 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:14:37 crc kubenswrapper[4675]: I0219 09:14:37.039021 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-pxgzv"] Feb 19 09:14:37 crc kubenswrapper[4675]: I0219 09:14:37.046944 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-pxgzv"] Feb 19 09:14:37 crc kubenswrapper[4675]: I0219 09:14:37.113424 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96629be8-e1e2-4cf3-8a9b-9df201341afb" path="/var/lib/kubelet/pods/96629be8-e1e2-4cf3-8a9b-9df201341afb/volumes" Feb 19 09:14:47 crc kubenswrapper[4675]: I0219 09:14:47.102864 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:14:47 crc kubenswrapper[4675]: E0219 09:14:47.103805 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:14:54 crc kubenswrapper[4675]: I0219 09:14:54.035826 4675 scope.go:117] "RemoveContainer" containerID="f5d7e5b8484626896ef08e590e60ad383e0ac383870f47f9316aca3e5a31e62e" Feb 19 09:14:54 crc kubenswrapper[4675]: I0219 09:14:54.081955 4675 scope.go:117] "RemoveContainer" containerID="3e3ce76b3c32778f16f95ff346231639cc54324dd25df20f79a0f043e8bd655a" Feb 19 09:14:59 crc kubenswrapper[4675]: I0219 09:14:59.109780 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:14:59 crc kubenswrapper[4675]: E0219 09:14:59.110802 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.148207 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6"] Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.149603 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.151848 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.152090 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.159153 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6"] Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.255976 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl4t7\" (UniqueName: \"kubernetes.io/projected/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-kube-api-access-nl4t7\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.256077 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-config-volume\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.256112 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-secret-volume\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.358890 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-secret-volume\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.359360 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl4t7\" (UniqueName: \"kubernetes.io/projected/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-kube-api-access-nl4t7\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.359420 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-config-volume\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.360352 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-config-volume\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.369177 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-secret-volume\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.378120 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl4t7\" (UniqueName: \"kubernetes.io/projected/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-kube-api-access-nl4t7\") pod \"collect-profiles-29524875-bwjh6\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.483537 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:00 crc kubenswrapper[4675]: I0219 09:15:00.930103 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6"] Feb 19 09:15:01 crc kubenswrapper[4675]: I0219 09:15:01.058122 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" event={"ID":"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f","Type":"ContainerStarted","Data":"327f1a95a3b635d4d76db53524ae7f5b1a40d652884a62046aa728c55ac3a7ae"} Feb 19 09:15:01 crc kubenswrapper[4675]: I0219 09:15:01.059869 4675 generic.go:334] "Generic (PLEG): container finished" podID="8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" containerID="69607d8b25f7db5590da483e8d9500965f55034699d379a6ccc28638e7b6f92d" exitCode=0 Feb 19 09:15:01 crc kubenswrapper[4675]: I0219 09:15:01.059904 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" event={"ID":"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa","Type":"ContainerDied","Data":"69607d8b25f7db5590da483e8d9500965f55034699d379a6ccc28638e7b6f92d"} Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.071491 4675 generic.go:334] "Generic (PLEG): container finished" podID="e08b2d97-6777-4fcf-9b54-2d7a00c7d01f" containerID="29eb2a0349cd2277005581dc23ad67e246acce00a38a4353998c87dfc5c6c429" exitCode=0 Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.071551 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" event={"ID":"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f","Type":"ContainerDied","Data":"29eb2a0349cd2277005581dc23ad67e246acce00a38a4353998c87dfc5c6c429"} Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.492276 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.609444 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.609507 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-inventory\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.609772 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ssh-key-openstack-edpm-ipam\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.610756 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-bootstrap-combined-ca-bundle\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.610855 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-telemetry-combined-ca-bundle\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.610904 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.610947 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46rb7\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-kube-api-access-46rb7\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.610993 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.611052 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-repo-setup-combined-ca-bundle\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.611087 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-nova-combined-ca-bundle\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.611136 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ovn-combined-ca-bundle\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.611180 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.611220 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-neutron-metadata-combined-ca-bundle\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.611278 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-libvirt-combined-ca-bundle\") pod \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\" (UID: \"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa\") " Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.618242 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.618596 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.618848 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.619405 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.619863 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.623488 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.623591 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.624617 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-kube-api-access-46rb7" (OuterVolumeSpecName: "kube-api-access-46rb7") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "kube-api-access-46rb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.624944 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.625312 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.625494 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.626988 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.649475 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-inventory" (OuterVolumeSpecName: "inventory") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.656028 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" (UID: "8dad42cd-cf8e-40d0-ab92-37c57c05d7fa"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.719789 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.720166 4675 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.720348 4675 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.720462 4675 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.720601 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46rb7\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-kube-api-access-46rb7\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.720712 4675 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.720903 4675 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.721037 4675 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.721119 4675 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.721191 4675 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.721261 4675 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.721339 4675 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.721418 4675 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:02 crc kubenswrapper[4675]: I0219 09:15:02.721504 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dad42cd-cf8e-40d0-ab92-37c57c05d7fa-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.082289 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.082315 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5cts8" event={"ID":"8dad42cd-cf8e-40d0-ab92-37c57c05d7fa","Type":"ContainerDied","Data":"925cacefbb8c679a07b4aa500ea1777b33d07236d2a45b66bcdaa30aa7701b6e"} Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.082392 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="925cacefbb8c679a07b4aa500ea1777b33d07236d2a45b66bcdaa30aa7701b6e" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.169001 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt"] Feb 19 09:15:03 crc kubenswrapper[4675]: E0219 09:15:03.169908 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.169931 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.170127 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dad42cd-cf8e-40d0-ab92-37c57c05d7fa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.170797 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.173020 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.173244 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.173399 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.174041 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.174859 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.186864 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt"] Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.335591 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.335917 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.336059 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5a7be99c-546f-4174-80ca-d72818c3ee43-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.336184 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.336320 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lkj2\" (UniqueName: \"kubernetes.io/projected/5a7be99c-546f-4174-80ca-d72818c3ee43-kube-api-access-7lkj2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.444231 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.444323 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.444380 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5a7be99c-546f-4174-80ca-d72818c3ee43-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.444431 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.444486 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lkj2\" (UniqueName: \"kubernetes.io/projected/5a7be99c-546f-4174-80ca-d72818c3ee43-kube-api-access-7lkj2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.446069 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5a7be99c-546f-4174-80ca-d72818c3ee43-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.450618 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.452082 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.457571 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.462001 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lkj2\" (UniqueName: \"kubernetes.io/projected/5a7be99c-546f-4174-80ca-d72818c3ee43-kube-api-access-7lkj2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cghvt\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.493906 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.622513 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.649792 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-config-volume\") pod \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.650160 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-secret-volume\") pod \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.650219 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl4t7\" (UniqueName: \"kubernetes.io/projected/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-kube-api-access-nl4t7\") pod \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\" (UID: \"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f\") " Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.651015 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-config-volume" (OuterVolumeSpecName: "config-volume") pod "e08b2d97-6777-4fcf-9b54-2d7a00c7d01f" (UID: "e08b2d97-6777-4fcf-9b54-2d7a00c7d01f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.657614 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e08b2d97-6777-4fcf-9b54-2d7a00c7d01f" (UID: "e08b2d97-6777-4fcf-9b54-2d7a00c7d01f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.660523 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-kube-api-access-nl4t7" (OuterVolumeSpecName: "kube-api-access-nl4t7") pod "e08b2d97-6777-4fcf-9b54-2d7a00c7d01f" (UID: "e08b2d97-6777-4fcf-9b54-2d7a00c7d01f"). InnerVolumeSpecName "kube-api-access-nl4t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.752248 4675 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-config-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.752652 4675 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:03 crc kubenswrapper[4675]: I0219 09:15:03.752664 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl4t7\" (UniqueName: \"kubernetes.io/projected/e08b2d97-6777-4fcf-9b54-2d7a00c7d01f-kube-api-access-nl4t7\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:04 crc kubenswrapper[4675]: I0219 09:15:04.037940 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt"] Feb 19 09:15:04 crc kubenswrapper[4675]: W0219 09:15:04.045994 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a7be99c_546f_4174_80ca_d72818c3ee43.slice/crio-7ebf62498775fccde7a5abf7153fd534ba46c92fcc042e17c969372f334ac3d7 WatchSource:0}: Error finding container 7ebf62498775fccde7a5abf7153fd534ba46c92fcc042e17c969372f334ac3d7: Status 404 returned error can't find the container with id 7ebf62498775fccde7a5abf7153fd534ba46c92fcc042e17c969372f334ac3d7 Feb 19 09:15:04 crc kubenswrapper[4675]: I0219 09:15:04.093144 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" event={"ID":"e08b2d97-6777-4fcf-9b54-2d7a00c7d01f","Type":"ContainerDied","Data":"327f1a95a3b635d4d76db53524ae7f5b1a40d652884a62046aa728c55ac3a7ae"} Feb 19 09:15:04 crc kubenswrapper[4675]: I0219 09:15:04.093201 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="327f1a95a3b635d4d76db53524ae7f5b1a40d652884a62046aa728c55ac3a7ae" Feb 19 09:15:04 crc kubenswrapper[4675]: I0219 09:15:04.093845 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524875-bwjh6" Feb 19 09:15:04 crc kubenswrapper[4675]: I0219 09:15:04.095439 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" event={"ID":"5a7be99c-546f-4174-80ca-d72818c3ee43","Type":"ContainerStarted","Data":"7ebf62498775fccde7a5abf7153fd534ba46c92fcc042e17c969372f334ac3d7"} Feb 19 09:15:04 crc kubenswrapper[4675]: E0219 09:15:04.291823 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode08b2d97_6777_4fcf_9b54_2d7a00c7d01f.slice/crio-327f1a95a3b635d4d76db53524ae7f5b1a40d652884a62046aa728c55ac3a7ae\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode08b2d97_6777_4fcf_9b54_2d7a00c7d01f.slice\": RecentStats: unable to find data in memory cache]" Feb 19 09:15:05 crc kubenswrapper[4675]: I0219 09:15:05.114559 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" event={"ID":"5a7be99c-546f-4174-80ca-d72818c3ee43","Type":"ContainerStarted","Data":"8a1cce41768b7655b896f268536b1e60b5e579a5065b36649f3c8615e0e809b7"} Feb 19 09:15:05 crc kubenswrapper[4675]: I0219 09:15:05.127397 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" podStartSLOduration=1.711885729 podStartE2EDuration="2.127377305s" podCreationTimestamp="2026-02-19 09:15:03 +0000 UTC" firstStartedPulling="2026-02-19 09:15:04.049368937 +0000 UTC m=+1885.676459205" lastFinishedPulling="2026-02-19 09:15:04.464860513 +0000 UTC m=+1886.091950781" observedRunningTime="2026-02-19 09:15:05.120312044 +0000 UTC m=+1886.747402302" watchObservedRunningTime="2026-02-19 09:15:05.127377305 +0000 UTC m=+1886.754467573" Feb 19 09:15:10 crc kubenswrapper[4675]: I0219 09:15:10.874841 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qch2b"] Feb 19 09:15:10 crc kubenswrapper[4675]: E0219 09:15:10.877005 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08b2d97-6777-4fcf-9b54-2d7a00c7d01f" containerName="collect-profiles" Feb 19 09:15:10 crc kubenswrapper[4675]: I0219 09:15:10.877024 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08b2d97-6777-4fcf-9b54-2d7a00c7d01f" containerName="collect-profiles" Feb 19 09:15:10 crc kubenswrapper[4675]: I0219 09:15:10.877192 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08b2d97-6777-4fcf-9b54-2d7a00c7d01f" containerName="collect-profiles" Feb 19 09:15:10 crc kubenswrapper[4675]: I0219 09:15:10.878704 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:10 crc kubenswrapper[4675]: I0219 09:15:10.888131 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qch2b"] Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.004058 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-catalog-content\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.004113 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-utilities\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.004143 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk5s4\" (UniqueName: \"kubernetes.io/projected/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-kube-api-access-bk5s4\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.105533 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-catalog-content\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.105614 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-utilities\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.105660 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk5s4\" (UniqueName: \"kubernetes.io/projected/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-kube-api-access-bk5s4\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.106360 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-catalog-content\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.106456 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-utilities\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.133271 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk5s4\" (UniqueName: \"kubernetes.io/projected/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-kube-api-access-bk5s4\") pod \"redhat-marketplace-qch2b\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.199144 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:11 crc kubenswrapper[4675]: I0219 09:15:11.752067 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qch2b"] Feb 19 09:15:12 crc kubenswrapper[4675]: I0219 09:15:12.165148 4675 generic.go:334] "Generic (PLEG): container finished" podID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerID="929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31" exitCode=0 Feb 19 09:15:12 crc kubenswrapper[4675]: I0219 09:15:12.165201 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qch2b" event={"ID":"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a","Type":"ContainerDied","Data":"929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31"} Feb 19 09:15:12 crc kubenswrapper[4675]: I0219 09:15:12.165234 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qch2b" event={"ID":"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a","Type":"ContainerStarted","Data":"0f0bea0f04242f249de43d174c5d13d2e9b6743dc752c9963cfe1a25b5f691fb"} Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.080887 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-62bbl"] Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.083242 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.097229 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62bbl"] Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.254873 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-catalog-content\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.254968 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-utilities\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.255109 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5hq4\" (UniqueName: \"kubernetes.io/projected/33ceaeb9-061c-454b-80af-7a7762b93bf3-kube-api-access-k5hq4\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.280470 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9nhs4"] Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.282734 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.301544 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9nhs4"] Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.357148 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5hq4\" (UniqueName: \"kubernetes.io/projected/33ceaeb9-061c-454b-80af-7a7762b93bf3-kube-api-access-k5hq4\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.357338 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-catalog-content\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.357367 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-utilities\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.357938 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-utilities\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.358040 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-catalog-content\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.381283 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5hq4\" (UniqueName: \"kubernetes.io/projected/33ceaeb9-061c-454b-80af-7a7762b93bf3-kube-api-access-k5hq4\") pod \"community-operators-62bbl\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.416684 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.459208 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-utilities\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.459298 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvlpc\" (UniqueName: \"kubernetes.io/projected/b657f21e-291a-4430-89f7-bc8ca97430ca-kube-api-access-jvlpc\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.459526 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-catalog-content\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.562183 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-utilities\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.562252 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvlpc\" (UniqueName: \"kubernetes.io/projected/b657f21e-291a-4430-89f7-bc8ca97430ca-kube-api-access-jvlpc\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.562342 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-catalog-content\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.567714 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-utilities\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.571595 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-catalog-content\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.593760 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvlpc\" (UniqueName: \"kubernetes.io/projected/b657f21e-291a-4430-89f7-bc8ca97430ca-kube-api-access-jvlpc\") pod \"certified-operators-9nhs4\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.606892 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:13 crc kubenswrapper[4675]: I0219 09:15:13.962066 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62bbl"] Feb 19 09:15:14 crc kubenswrapper[4675]: I0219 09:15:14.104241 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:15:14 crc kubenswrapper[4675]: E0219 09:15:14.104968 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:15:14 crc kubenswrapper[4675]: I0219 09:15:14.202039 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9nhs4"] Feb 19 09:15:14 crc kubenswrapper[4675]: I0219 09:15:14.206163 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62bbl" event={"ID":"33ceaeb9-061c-454b-80af-7a7762b93bf3","Type":"ContainerStarted","Data":"6cbd6c5b9997debfde36781d7190cedf249cb8ce89e0a814dbc780d4ef0c2940"} Feb 19 09:15:14 crc kubenswrapper[4675]: I0219 09:15:14.209046 4675 generic.go:334] "Generic (PLEG): container finished" podID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerID="65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267" exitCode=0 Feb 19 09:15:14 crc kubenswrapper[4675]: I0219 09:15:14.209093 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qch2b" event={"ID":"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a","Type":"ContainerDied","Data":"65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267"} Feb 19 09:15:14 crc kubenswrapper[4675]: W0219 09:15:14.225693 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb657f21e_291a_4430_89f7_bc8ca97430ca.slice/crio-f868a15e9740320fedd8aa7b41695c49ee16defdf8085845264abb1915b453d0 WatchSource:0}: Error finding container f868a15e9740320fedd8aa7b41695c49ee16defdf8085845264abb1915b453d0: Status 404 returned error can't find the container with id f868a15e9740320fedd8aa7b41695c49ee16defdf8085845264abb1915b453d0 Feb 19 09:15:15 crc kubenswrapper[4675]: I0219 09:15:15.221071 4675 generic.go:334] "Generic (PLEG): container finished" podID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerID="54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d" exitCode=0 Feb 19 09:15:15 crc kubenswrapper[4675]: I0219 09:15:15.221395 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62bbl" event={"ID":"33ceaeb9-061c-454b-80af-7a7762b93bf3","Type":"ContainerDied","Data":"54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d"} Feb 19 09:15:15 crc kubenswrapper[4675]: I0219 09:15:15.226777 4675 generic.go:334] "Generic (PLEG): container finished" podID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerID="cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17" exitCode=0 Feb 19 09:15:15 crc kubenswrapper[4675]: I0219 09:15:15.226990 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9nhs4" event={"ID":"b657f21e-291a-4430-89f7-bc8ca97430ca","Type":"ContainerDied","Data":"cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17"} Feb 19 09:15:15 crc kubenswrapper[4675]: I0219 09:15:15.227288 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9nhs4" event={"ID":"b657f21e-291a-4430-89f7-bc8ca97430ca","Type":"ContainerStarted","Data":"f868a15e9740320fedd8aa7b41695c49ee16defdf8085845264abb1915b453d0"} Feb 19 09:15:15 crc kubenswrapper[4675]: I0219 09:15:15.233997 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qch2b" event={"ID":"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a","Type":"ContainerStarted","Data":"22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43"} Feb 19 09:15:15 crc kubenswrapper[4675]: I0219 09:15:15.277567 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qch2b" podStartSLOduration=2.775129523 podStartE2EDuration="5.277537333s" podCreationTimestamp="2026-02-19 09:15:10 +0000 UTC" firstStartedPulling="2026-02-19 09:15:12.167045864 +0000 UTC m=+1893.794136132" lastFinishedPulling="2026-02-19 09:15:14.669453674 +0000 UTC m=+1896.296543942" observedRunningTime="2026-02-19 09:15:15.2703867 +0000 UTC m=+1896.897476978" watchObservedRunningTime="2026-02-19 09:15:15.277537333 +0000 UTC m=+1896.904627601" Feb 19 09:15:16 crc kubenswrapper[4675]: I0219 09:15:16.249864 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62bbl" event={"ID":"33ceaeb9-061c-454b-80af-7a7762b93bf3","Type":"ContainerStarted","Data":"78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559"} Feb 19 09:15:16 crc kubenswrapper[4675]: I0219 09:15:16.269732 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9nhs4" event={"ID":"b657f21e-291a-4430-89f7-bc8ca97430ca","Type":"ContainerStarted","Data":"099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca"} Feb 19 09:15:17 crc kubenswrapper[4675]: I0219 09:15:17.278558 4675 generic.go:334] "Generic (PLEG): container finished" podID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerID="78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559" exitCode=0 Feb 19 09:15:17 crc kubenswrapper[4675]: I0219 09:15:17.278648 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62bbl" event={"ID":"33ceaeb9-061c-454b-80af-7a7762b93bf3","Type":"ContainerDied","Data":"78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559"} Feb 19 09:15:17 crc kubenswrapper[4675]: I0219 09:15:17.282245 4675 generic.go:334] "Generic (PLEG): container finished" podID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerID="099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca" exitCode=0 Feb 19 09:15:17 crc kubenswrapper[4675]: I0219 09:15:17.282292 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9nhs4" event={"ID":"b657f21e-291a-4430-89f7-bc8ca97430ca","Type":"ContainerDied","Data":"099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca"} Feb 19 09:15:17 crc kubenswrapper[4675]: I0219 09:15:17.282336 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9nhs4" event={"ID":"b657f21e-291a-4430-89f7-bc8ca97430ca","Type":"ContainerStarted","Data":"eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51"} Feb 19 09:15:17 crc kubenswrapper[4675]: I0219 09:15:17.320030 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9nhs4" podStartSLOduration=2.893518226 podStartE2EDuration="4.320005205s" podCreationTimestamp="2026-02-19 09:15:13 +0000 UTC" firstStartedPulling="2026-02-19 09:15:15.228571828 +0000 UTC m=+1896.855662096" lastFinishedPulling="2026-02-19 09:15:16.655058807 +0000 UTC m=+1898.282149075" observedRunningTime="2026-02-19 09:15:17.317112647 +0000 UTC m=+1898.944202915" watchObservedRunningTime="2026-02-19 09:15:17.320005205 +0000 UTC m=+1898.947095473" Feb 19 09:15:18 crc kubenswrapper[4675]: I0219 09:15:18.295074 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62bbl" event={"ID":"33ceaeb9-061c-454b-80af-7a7762b93bf3","Type":"ContainerStarted","Data":"cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff"} Feb 19 09:15:18 crc kubenswrapper[4675]: I0219 09:15:18.323470 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-62bbl" podStartSLOduration=2.878318764 podStartE2EDuration="5.323449525s" podCreationTimestamp="2026-02-19 09:15:13 +0000 UTC" firstStartedPulling="2026-02-19 09:15:15.223599253 +0000 UTC m=+1896.850689521" lastFinishedPulling="2026-02-19 09:15:17.668730014 +0000 UTC m=+1899.295820282" observedRunningTime="2026-02-19 09:15:18.318042668 +0000 UTC m=+1899.945132956" watchObservedRunningTime="2026-02-19 09:15:18.323449525 +0000 UTC m=+1899.950539793" Feb 19 09:15:21 crc kubenswrapper[4675]: I0219 09:15:21.200345 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:21 crc kubenswrapper[4675]: I0219 09:15:21.201679 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:21 crc kubenswrapper[4675]: I0219 09:15:21.245510 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:21 crc kubenswrapper[4675]: I0219 09:15:21.411672 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:23 crc kubenswrapper[4675]: I0219 09:15:23.417300 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:23 crc kubenswrapper[4675]: I0219 09:15:23.417775 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:23 crc kubenswrapper[4675]: I0219 09:15:23.463736 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:23 crc kubenswrapper[4675]: I0219 09:15:23.607382 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:23 crc kubenswrapper[4675]: I0219 09:15:23.607434 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:23 crc kubenswrapper[4675]: I0219 09:15:23.659276 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:24 crc kubenswrapper[4675]: I0219 09:15:24.388622 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:24 crc kubenswrapper[4675]: I0219 09:15:24.396402 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:25 crc kubenswrapper[4675]: I0219 09:15:25.862652 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qch2b"] Feb 19 09:15:25 crc kubenswrapper[4675]: I0219 09:15:25.863622 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qch2b" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="registry-server" containerID="cri-o://22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43" gracePeriod=2 Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.397693 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.398656 4675 generic.go:334] "Generic (PLEG): container finished" podID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerID="22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43" exitCode=0 Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.398725 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qch2b" event={"ID":"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a","Type":"ContainerDied","Data":"22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43"} Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.398805 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qch2b" event={"ID":"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a","Type":"ContainerDied","Data":"0f0bea0f04242f249de43d174c5d13d2e9b6743dc752c9963cfe1a25b5f691fb"} Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.398858 4675 scope.go:117] "RemoveContainer" containerID="22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.423806 4675 scope.go:117] "RemoveContainer" containerID="65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.457931 4675 scope.go:117] "RemoveContainer" containerID="929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.513299 4675 scope.go:117] "RemoveContainer" containerID="22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43" Feb 19 09:15:26 crc kubenswrapper[4675]: E0219 09:15:26.513899 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43\": container with ID starting with 22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43 not found: ID does not exist" containerID="22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.513943 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43"} err="failed to get container status \"22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43\": rpc error: code = NotFound desc = could not find container \"22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43\": container with ID starting with 22728d93702813845ae64337b1778bf85a5e9351120344ef3bd680d961aa1c43 not found: ID does not exist" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.513970 4675 scope.go:117] "RemoveContainer" containerID="65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267" Feb 19 09:15:26 crc kubenswrapper[4675]: E0219 09:15:26.514196 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267\": container with ID starting with 65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267 not found: ID does not exist" containerID="65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.514221 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267"} err="failed to get container status \"65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267\": rpc error: code = NotFound desc = could not find container \"65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267\": container with ID starting with 65ad548f84fd5b6597fae8819ac71f3c123e3045ac889fcda1682635fffd3267 not found: ID does not exist" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.514237 4675 scope.go:117] "RemoveContainer" containerID="929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31" Feb 19 09:15:26 crc kubenswrapper[4675]: E0219 09:15:26.514462 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31\": container with ID starting with 929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31 not found: ID does not exist" containerID="929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.514484 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31"} err="failed to get container status \"929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31\": rpc error: code = NotFound desc = could not find container \"929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31\": container with ID starting with 929ca7168a87c3bd3208dcc0e55e6fd99ae18643d7142a74ff2cdd5a2cbdeb31 not found: ID does not exist" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.588037 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk5s4\" (UniqueName: \"kubernetes.io/projected/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-kube-api-access-bk5s4\") pod \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.588288 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-catalog-content\") pod \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.588673 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-utilities\") pod \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\" (UID: \"66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a\") " Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.589603 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-utilities" (OuterVolumeSpecName: "utilities") pod "66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" (UID: "66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.594732 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-kube-api-access-bk5s4" (OuterVolumeSpecName: "kube-api-access-bk5s4") pod "66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" (UID: "66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a"). InnerVolumeSpecName "kube-api-access-bk5s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.610277 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" (UID: "66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.690959 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.690994 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk5s4\" (UniqueName: \"kubernetes.io/projected/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-kube-api-access-bk5s4\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:26 crc kubenswrapper[4675]: I0219 09:15:26.691005 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:27 crc kubenswrapper[4675]: I0219 09:15:27.409746 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qch2b" Feb 19 09:15:27 crc kubenswrapper[4675]: I0219 09:15:27.440241 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qch2b"] Feb 19 09:15:27 crc kubenswrapper[4675]: I0219 09:15:27.450916 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qch2b"] Feb 19 09:15:27 crc kubenswrapper[4675]: I0219 09:15:27.864030 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9nhs4"] Feb 19 09:15:27 crc kubenswrapper[4675]: I0219 09:15:27.864766 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9nhs4" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="registry-server" containerID="cri-o://eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51" gracePeriod=2 Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.327418 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.418893 4675 generic.go:334] "Generic (PLEG): container finished" podID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerID="eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51" exitCode=0 Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.418949 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9nhs4" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.418997 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9nhs4" event={"ID":"b657f21e-291a-4430-89f7-bc8ca97430ca","Type":"ContainerDied","Data":"eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51"} Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.420062 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9nhs4" event={"ID":"b657f21e-291a-4430-89f7-bc8ca97430ca","Type":"ContainerDied","Data":"f868a15e9740320fedd8aa7b41695c49ee16defdf8085845264abb1915b453d0"} Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.420098 4675 scope.go:117] "RemoveContainer" containerID="eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.426720 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-catalog-content\") pod \"b657f21e-291a-4430-89f7-bc8ca97430ca\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.426776 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvlpc\" (UniqueName: \"kubernetes.io/projected/b657f21e-291a-4430-89f7-bc8ca97430ca-kube-api-access-jvlpc\") pod \"b657f21e-291a-4430-89f7-bc8ca97430ca\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.427022 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-utilities\") pod \"b657f21e-291a-4430-89f7-bc8ca97430ca\" (UID: \"b657f21e-291a-4430-89f7-bc8ca97430ca\") " Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.428166 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-utilities" (OuterVolumeSpecName: "utilities") pod "b657f21e-291a-4430-89f7-bc8ca97430ca" (UID: "b657f21e-291a-4430-89f7-bc8ca97430ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.433915 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b657f21e-291a-4430-89f7-bc8ca97430ca-kube-api-access-jvlpc" (OuterVolumeSpecName: "kube-api-access-jvlpc") pod "b657f21e-291a-4430-89f7-bc8ca97430ca" (UID: "b657f21e-291a-4430-89f7-bc8ca97430ca"). InnerVolumeSpecName "kube-api-access-jvlpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.441192 4675 scope.go:117] "RemoveContainer" containerID="099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.479868 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b657f21e-291a-4430-89f7-bc8ca97430ca" (UID: "b657f21e-291a-4430-89f7-bc8ca97430ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.499372 4675 scope.go:117] "RemoveContainer" containerID="cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.530115 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.530159 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvlpc\" (UniqueName: \"kubernetes.io/projected/b657f21e-291a-4430-89f7-bc8ca97430ca-kube-api-access-jvlpc\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.530172 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b657f21e-291a-4430-89f7-bc8ca97430ca-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.546864 4675 scope.go:117] "RemoveContainer" containerID="eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51" Feb 19 09:15:28 crc kubenswrapper[4675]: E0219 09:15:28.547476 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51\": container with ID starting with eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51 not found: ID does not exist" containerID="eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.547537 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51"} err="failed to get container status \"eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51\": rpc error: code = NotFound desc = could not find container \"eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51\": container with ID starting with eeb2ef7a180293b6b1013c414843fe266633d93471ba0b88e47da590968c3a51 not found: ID does not exist" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.547569 4675 scope.go:117] "RemoveContainer" containerID="099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca" Feb 19 09:15:28 crc kubenswrapper[4675]: E0219 09:15:28.548081 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca\": container with ID starting with 099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca not found: ID does not exist" containerID="099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.548121 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca"} err="failed to get container status \"099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca\": rpc error: code = NotFound desc = could not find container \"099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca\": container with ID starting with 099045502fa2f1a905f368b1d065262771fb3aa3cad03dc3a9e6be49f63b6dca not found: ID does not exist" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.548149 4675 scope.go:117] "RemoveContainer" containerID="cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17" Feb 19 09:15:28 crc kubenswrapper[4675]: E0219 09:15:28.548691 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17\": container with ID starting with cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17 not found: ID does not exist" containerID="cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.548721 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17"} err="failed to get container status \"cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17\": rpc error: code = NotFound desc = could not find container \"cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17\": container with ID starting with cccdd75b5035e630e4b3749e8cb04647ea157e6db06643e19847b88691085d17 not found: ID does not exist" Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.755656 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9nhs4"] Feb 19 09:15:28 crc kubenswrapper[4675]: I0219 09:15:28.763469 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9nhs4"] Feb 19 09:15:29 crc kubenswrapper[4675]: I0219 09:15:29.109333 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:15:29 crc kubenswrapper[4675]: E0219 09:15:29.109685 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:15:29 crc kubenswrapper[4675]: I0219 09:15:29.113463 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" path="/var/lib/kubelet/pods/66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a/volumes" Feb 19 09:15:29 crc kubenswrapper[4675]: I0219 09:15:29.114299 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" path="/var/lib/kubelet/pods/b657f21e-291a-4430-89f7-bc8ca97430ca/volumes" Feb 19 09:15:29 crc kubenswrapper[4675]: I0219 09:15:29.462440 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62bbl"] Feb 19 09:15:29 crc kubenswrapper[4675]: I0219 09:15:29.462717 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-62bbl" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="registry-server" containerID="cri-o://cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff" gracePeriod=2 Feb 19 09:15:29 crc kubenswrapper[4675]: I0219 09:15:29.932227 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.057957 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-utilities\") pod \"33ceaeb9-061c-454b-80af-7a7762b93bf3\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.058405 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5hq4\" (UniqueName: \"kubernetes.io/projected/33ceaeb9-061c-454b-80af-7a7762b93bf3-kube-api-access-k5hq4\") pod \"33ceaeb9-061c-454b-80af-7a7762b93bf3\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.058477 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-catalog-content\") pod \"33ceaeb9-061c-454b-80af-7a7762b93bf3\" (UID: \"33ceaeb9-061c-454b-80af-7a7762b93bf3\") " Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.059606 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-utilities" (OuterVolumeSpecName: "utilities") pod "33ceaeb9-061c-454b-80af-7a7762b93bf3" (UID: "33ceaeb9-061c-454b-80af-7a7762b93bf3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.066849 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33ceaeb9-061c-454b-80af-7a7762b93bf3-kube-api-access-k5hq4" (OuterVolumeSpecName: "kube-api-access-k5hq4") pod "33ceaeb9-061c-454b-80af-7a7762b93bf3" (UID: "33ceaeb9-061c-454b-80af-7a7762b93bf3"). InnerVolumeSpecName "kube-api-access-k5hq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.114867 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33ceaeb9-061c-454b-80af-7a7762b93bf3" (UID: "33ceaeb9-061c-454b-80af-7a7762b93bf3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.160363 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.160402 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ceaeb9-061c-454b-80af-7a7762b93bf3-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.160421 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5hq4\" (UniqueName: \"kubernetes.io/projected/33ceaeb9-061c-454b-80af-7a7762b93bf3-kube-api-access-k5hq4\") on node \"crc\" DevicePath \"\"" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.438557 4675 generic.go:334] "Generic (PLEG): container finished" podID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerID="cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff" exitCode=0 Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.438604 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62bbl" event={"ID":"33ceaeb9-061c-454b-80af-7a7762b93bf3","Type":"ContainerDied","Data":"cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff"} Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.438632 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62bbl" event={"ID":"33ceaeb9-061c-454b-80af-7a7762b93bf3","Type":"ContainerDied","Data":"6cbd6c5b9997debfde36781d7190cedf249cb8ce89e0a814dbc780d4ef0c2940"} Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.438663 4675 scope.go:117] "RemoveContainer" containerID="cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.438764 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62bbl" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.468524 4675 scope.go:117] "RemoveContainer" containerID="78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.484953 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62bbl"] Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.494986 4675 scope.go:117] "RemoveContainer" containerID="54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.495758 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-62bbl"] Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.536075 4675 scope.go:117] "RemoveContainer" containerID="cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff" Feb 19 09:15:30 crc kubenswrapper[4675]: E0219 09:15:30.536593 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff\": container with ID starting with cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff not found: ID does not exist" containerID="cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.536665 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff"} err="failed to get container status \"cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff\": rpc error: code = NotFound desc = could not find container \"cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff\": container with ID starting with cacd00649b5813d07b8f752b7f37cbefe91b9f51873f2bbccd7e2f42f3a059ff not found: ID does not exist" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.536704 4675 scope.go:117] "RemoveContainer" containerID="78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559" Feb 19 09:15:30 crc kubenswrapper[4675]: E0219 09:15:30.537023 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559\": container with ID starting with 78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559 not found: ID does not exist" containerID="78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.537057 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559"} err="failed to get container status \"78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559\": rpc error: code = NotFound desc = could not find container \"78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559\": container with ID starting with 78e101ae4d580d550675bcca43001b54003d65ae1c951ff6354e59ea6ef20559 not found: ID does not exist" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.537081 4675 scope.go:117] "RemoveContainer" containerID="54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d" Feb 19 09:15:30 crc kubenswrapper[4675]: E0219 09:15:30.537404 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d\": container with ID starting with 54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d not found: ID does not exist" containerID="54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d" Feb 19 09:15:30 crc kubenswrapper[4675]: I0219 09:15:30.537438 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d"} err="failed to get container status \"54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d\": rpc error: code = NotFound desc = could not find container \"54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d\": container with ID starting with 54e27d54ee497fd64f7026f419ca63b3d8715bbe33d3f14cf4d20fe7f086148d not found: ID does not exist" Feb 19 09:15:31 crc kubenswrapper[4675]: I0219 09:15:31.129607 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" path="/var/lib/kubelet/pods/33ceaeb9-061c-454b-80af-7a7762b93bf3/volumes" Feb 19 09:15:40 crc kubenswrapper[4675]: I0219 09:15:40.103865 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:15:40 crc kubenswrapper[4675]: E0219 09:15:40.104774 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:15:54 crc kubenswrapper[4675]: I0219 09:15:54.103681 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:15:54 crc kubenswrapper[4675]: E0219 09:15:54.104700 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:16:02 crc kubenswrapper[4675]: I0219 09:16:02.708283 4675 generic.go:334] "Generic (PLEG): container finished" podID="5a7be99c-546f-4174-80ca-d72818c3ee43" containerID="8a1cce41768b7655b896f268536b1e60b5e579a5065b36649f3c8615e0e809b7" exitCode=0 Feb 19 09:16:02 crc kubenswrapper[4675]: I0219 09:16:02.708501 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" event={"ID":"5a7be99c-546f-4174-80ca-d72818c3ee43","Type":"ContainerDied","Data":"8a1cce41768b7655b896f268536b1e60b5e579a5065b36649f3c8615e0e809b7"} Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.163095 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.332573 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ovn-combined-ca-bundle\") pod \"5a7be99c-546f-4174-80ca-d72818c3ee43\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.332714 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ssh-key-openstack-edpm-ipam\") pod \"5a7be99c-546f-4174-80ca-d72818c3ee43\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.332843 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5a7be99c-546f-4174-80ca-d72818c3ee43-ovncontroller-config-0\") pod \"5a7be99c-546f-4174-80ca-d72818c3ee43\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.332932 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lkj2\" (UniqueName: \"kubernetes.io/projected/5a7be99c-546f-4174-80ca-d72818c3ee43-kube-api-access-7lkj2\") pod \"5a7be99c-546f-4174-80ca-d72818c3ee43\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.333028 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-inventory\") pod \"5a7be99c-546f-4174-80ca-d72818c3ee43\" (UID: \"5a7be99c-546f-4174-80ca-d72818c3ee43\") " Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.340068 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a7be99c-546f-4174-80ca-d72818c3ee43-kube-api-access-7lkj2" (OuterVolumeSpecName: "kube-api-access-7lkj2") pod "5a7be99c-546f-4174-80ca-d72818c3ee43" (UID: "5a7be99c-546f-4174-80ca-d72818c3ee43"). InnerVolumeSpecName "kube-api-access-7lkj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.346978 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5a7be99c-546f-4174-80ca-d72818c3ee43" (UID: "5a7be99c-546f-4174-80ca-d72818c3ee43"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.360157 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a7be99c-546f-4174-80ca-d72818c3ee43-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "5a7be99c-546f-4174-80ca-d72818c3ee43" (UID: "5a7be99c-546f-4174-80ca-d72818c3ee43"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.363809 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-inventory" (OuterVolumeSpecName: "inventory") pod "5a7be99c-546f-4174-80ca-d72818c3ee43" (UID: "5a7be99c-546f-4174-80ca-d72818c3ee43"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.364916 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5a7be99c-546f-4174-80ca-d72818c3ee43" (UID: "5a7be99c-546f-4174-80ca-d72818c3ee43"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.435503 4675 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5a7be99c-546f-4174-80ca-d72818c3ee43-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.435554 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lkj2\" (UniqueName: \"kubernetes.io/projected/5a7be99c-546f-4174-80ca-d72818c3ee43-kube-api-access-7lkj2\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.435568 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.435581 4675 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.435596 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5a7be99c-546f-4174-80ca-d72818c3ee43-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.728468 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" event={"ID":"5a7be99c-546f-4174-80ca-d72818c3ee43","Type":"ContainerDied","Data":"7ebf62498775fccde7a5abf7153fd534ba46c92fcc042e17c969372f334ac3d7"} Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.728566 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ebf62498775fccde7a5abf7153fd534ba46c92fcc042e17c969372f334ac3d7" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.728565 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cghvt" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.830840 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9"] Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831364 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="extract-utilities" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831391 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="extract-utilities" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831409 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831416 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831425 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="extract-utilities" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831436 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="extract-utilities" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831460 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831469 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831484 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="extract-utilities" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831491 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="extract-utilities" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831511 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="extract-content" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831521 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="extract-content" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831533 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="extract-content" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831540 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="extract-content" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831560 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="extract-content" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831567 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="extract-content" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831576 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831583 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: E0219 09:16:04.831603 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a7be99c-546f-4174-80ca-d72818c3ee43" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831610 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a7be99c-546f-4174-80ca-d72818c3ee43" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831842 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="33ceaeb9-061c-454b-80af-7a7762b93bf3" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831865 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b45031-29cf-4dd1-87c1-8eb2ab5a5a9a" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831887 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="b657f21e-291a-4430-89f7-bc8ca97430ca" containerName="registry-server" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.831901 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a7be99c-546f-4174-80ca-d72818c3ee43" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.835162 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.838162 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.838405 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.838577 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.839184 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.839570 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.843503 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9"] Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.843713 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.843812 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.843965 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkrww\" (UniqueName: \"kubernetes.io/projected/185f89ae-500c-4bdf-81b1-e3627ab1d600-kube-api-access-jkrww\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.844290 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.844369 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.844425 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.857937 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.946687 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.947189 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.947223 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.947286 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.947332 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.947397 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkrww\" (UniqueName: \"kubernetes.io/projected/185f89ae-500c-4bdf-81b1-e3627ab1d600-kube-api-access-jkrww\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.953035 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.953075 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.953605 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.954549 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.955260 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:04 crc kubenswrapper[4675]: I0219 09:16:04.966656 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkrww\" (UniqueName: \"kubernetes.io/projected/185f89ae-500c-4bdf-81b1-e3627ab1d600-kube-api-access-jkrww\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:05 crc kubenswrapper[4675]: I0219 09:16:05.164384 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:05 crc kubenswrapper[4675]: I0219 09:16:05.928857 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9"] Feb 19 09:16:06 crc kubenswrapper[4675]: I0219 09:16:06.103693 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:16:06 crc kubenswrapper[4675]: E0219 09:16:06.104691 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:16:06 crc kubenswrapper[4675]: I0219 09:16:06.821141 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" event={"ID":"185f89ae-500c-4bdf-81b1-e3627ab1d600","Type":"ContainerStarted","Data":"805529db59ced235f459c591d45bb48fd3221706601025790290132b5a206f89"} Feb 19 09:16:06 crc kubenswrapper[4675]: I0219 09:16:06.821605 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" event={"ID":"185f89ae-500c-4bdf-81b1-e3627ab1d600","Type":"ContainerStarted","Data":"85ebe87525b28062819461e8f23dfba8b37e942f50bdec2bdda7ffe76bb55a6c"} Feb 19 09:16:06 crc kubenswrapper[4675]: I0219 09:16:06.845314 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" podStartSLOduration=2.40026131 podStartE2EDuration="2.845295956s" podCreationTimestamp="2026-02-19 09:16:04 +0000 UTC" firstStartedPulling="2026-02-19 09:16:05.937578288 +0000 UTC m=+1947.564668556" lastFinishedPulling="2026-02-19 09:16:06.382612934 +0000 UTC m=+1948.009703202" observedRunningTime="2026-02-19 09:16:06.835736227 +0000 UTC m=+1948.462826505" watchObservedRunningTime="2026-02-19 09:16:06.845295956 +0000 UTC m=+1948.472386224" Feb 19 09:16:18 crc kubenswrapper[4675]: I0219 09:16:18.103938 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:16:18 crc kubenswrapper[4675]: E0219 09:16:18.104935 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:16:29 crc kubenswrapper[4675]: I0219 09:16:29.110155 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:16:29 crc kubenswrapper[4675]: E0219 09:16:29.111285 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:16:43 crc kubenswrapper[4675]: I0219 09:16:43.103262 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:16:44 crc kubenswrapper[4675]: I0219 09:16:44.142220 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293"} Feb 19 09:16:51 crc kubenswrapper[4675]: I0219 09:16:51.202662 4675 generic.go:334] "Generic (PLEG): container finished" podID="185f89ae-500c-4bdf-81b1-e3627ab1d600" containerID="805529db59ced235f459c591d45bb48fd3221706601025790290132b5a206f89" exitCode=0 Feb 19 09:16:51 crc kubenswrapper[4675]: I0219 09:16:51.202739 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" event={"ID":"185f89ae-500c-4bdf-81b1-e3627ab1d600","Type":"ContainerDied","Data":"805529db59ced235f459c591d45bb48fd3221706601025790290132b5a206f89"} Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.617427 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.705574 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-inventory\") pod \"185f89ae-500c-4bdf-81b1-e3627ab1d600\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.705750 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-ssh-key-openstack-edpm-ipam\") pod \"185f89ae-500c-4bdf-81b1-e3627ab1d600\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.705812 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-ovn-metadata-agent-neutron-config-0\") pod \"185f89ae-500c-4bdf-81b1-e3627ab1d600\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.705892 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-metadata-combined-ca-bundle\") pod \"185f89ae-500c-4bdf-81b1-e3627ab1d600\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.705977 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkrww\" (UniqueName: \"kubernetes.io/projected/185f89ae-500c-4bdf-81b1-e3627ab1d600-kube-api-access-jkrww\") pod \"185f89ae-500c-4bdf-81b1-e3627ab1d600\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.706041 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-nova-metadata-neutron-config-0\") pod \"185f89ae-500c-4bdf-81b1-e3627ab1d600\" (UID: \"185f89ae-500c-4bdf-81b1-e3627ab1d600\") " Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.712329 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "185f89ae-500c-4bdf-81b1-e3627ab1d600" (UID: "185f89ae-500c-4bdf-81b1-e3627ab1d600"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.712539 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185f89ae-500c-4bdf-81b1-e3627ab1d600-kube-api-access-jkrww" (OuterVolumeSpecName: "kube-api-access-jkrww") pod "185f89ae-500c-4bdf-81b1-e3627ab1d600" (UID: "185f89ae-500c-4bdf-81b1-e3627ab1d600"). InnerVolumeSpecName "kube-api-access-jkrww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.733512 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "185f89ae-500c-4bdf-81b1-e3627ab1d600" (UID: "185f89ae-500c-4bdf-81b1-e3627ab1d600"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.734822 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "185f89ae-500c-4bdf-81b1-e3627ab1d600" (UID: "185f89ae-500c-4bdf-81b1-e3627ab1d600"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.735484 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "185f89ae-500c-4bdf-81b1-e3627ab1d600" (UID: "185f89ae-500c-4bdf-81b1-e3627ab1d600"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.735950 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-inventory" (OuterVolumeSpecName: "inventory") pod "185f89ae-500c-4bdf-81b1-e3627ab1d600" (UID: "185f89ae-500c-4bdf-81b1-e3627ab1d600"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.809535 4675 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.809868 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.809979 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.810060 4675 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.810139 4675 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f89ae-500c-4bdf-81b1-e3627ab1d600-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:52 crc kubenswrapper[4675]: I0219 09:16:52.810312 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkrww\" (UniqueName: \"kubernetes.io/projected/185f89ae-500c-4bdf-81b1-e3627ab1d600-kube-api-access-jkrww\") on node \"crc\" DevicePath \"\"" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.220397 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" event={"ID":"185f89ae-500c-4bdf-81b1-e3627ab1d600","Type":"ContainerDied","Data":"85ebe87525b28062819461e8f23dfba8b37e942f50bdec2bdda7ffe76bb55a6c"} Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.220857 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ebe87525b28062819461e8f23dfba8b37e942f50bdec2bdda7ffe76bb55a6c" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.220465 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.316431 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk"] Feb 19 09:16:53 crc kubenswrapper[4675]: E0219 09:16:53.316983 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185f89ae-500c-4bdf-81b1-e3627ab1d600" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.317008 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="185f89ae-500c-4bdf-81b1-e3627ab1d600" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.317262 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="185f89ae-500c-4bdf-81b1-e3627ab1d600" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.318217 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.321031 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.321325 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.323191 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.326495 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.329459 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk"] Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.329730 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.420796 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.420864 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/12901b5c-afde-4017-86ba-363d50c78a9a-kube-api-access-j6sht\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.420914 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.421067 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.421308 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.523183 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.523270 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.523330 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.523367 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/12901b5c-afde-4017-86ba-363d50c78a9a-kube-api-access-j6sht\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.523416 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.529703 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.529727 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.530181 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.531066 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.544920 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/12901b5c-afde-4017-86ba-363d50c78a9a-kube-api-access-j6sht\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pknwk\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:53 crc kubenswrapper[4675]: I0219 09:16:53.634825 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:16:54 crc kubenswrapper[4675]: I0219 09:16:54.177002 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk"] Feb 19 09:16:54 crc kubenswrapper[4675]: I0219 09:16:54.192571 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:16:54 crc kubenswrapper[4675]: I0219 09:16:54.229946 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" event={"ID":"12901b5c-afde-4017-86ba-363d50c78a9a","Type":"ContainerStarted","Data":"cce3cfa5f310c253fb711554a2a74477ea7633e1a0e6fcd1430108ea48d23b67"} Feb 19 09:16:55 crc kubenswrapper[4675]: I0219 09:16:55.239174 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" event={"ID":"12901b5c-afde-4017-86ba-363d50c78a9a","Type":"ContainerStarted","Data":"02a77cc089ac5389bd7e96e4434f2e59012b0743a1fa26d5709fd42f1d549923"} Feb 19 09:16:55 crc kubenswrapper[4675]: I0219 09:16:55.263427 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" podStartSLOduration=1.768432163 podStartE2EDuration="2.26340683s" podCreationTimestamp="2026-02-19 09:16:53 +0000 UTC" firstStartedPulling="2026-02-19 09:16:54.192389411 +0000 UTC m=+1995.819479679" lastFinishedPulling="2026-02-19 09:16:54.687364078 +0000 UTC m=+1996.314454346" observedRunningTime="2026-02-19 09:16:55.254698554 +0000 UTC m=+1996.881788842" watchObservedRunningTime="2026-02-19 09:16:55.26340683 +0000 UTC m=+1996.890497098" Feb 19 09:19:11 crc kubenswrapper[4675]: I0219 09:19:11.743582 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:19:11 crc kubenswrapper[4675]: I0219 09:19:11.744474 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:19:41 crc kubenswrapper[4675]: I0219 09:19:41.743487 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:19:41 crc kubenswrapper[4675]: I0219 09:19:41.744452 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.743569 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.744514 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.744570 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.745670 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.745739 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293" gracePeriod=600 Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.997470 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293" exitCode=0 Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.997531 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293"} Feb 19 09:20:11 crc kubenswrapper[4675]: I0219 09:20:11.997579 4675 scope.go:117] "RemoveContainer" containerID="2bd2210b9553cce30f1d9c9c69e356e12861ff8dc1d7c48150345afc9500cf44" Feb 19 09:20:12 crc kubenswrapper[4675]: E0219 09:20:12.002810 4675 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e91469d_12b7_434c_991d_633287712a69.slice/crio-conmon-e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e91469d_12b7_434c_991d_633287712a69.slice/crio-e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293.scope\": RecentStats: unable to find data in memory cache]" Feb 19 09:20:13 crc kubenswrapper[4675]: I0219 09:20:13.008874 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b"} Feb 19 09:20:25 crc kubenswrapper[4675]: I0219 09:20:25.111180 4675 generic.go:334] "Generic (PLEG): container finished" podID="12901b5c-afde-4017-86ba-363d50c78a9a" containerID="02a77cc089ac5389bd7e96e4434f2e59012b0743a1fa26d5709fd42f1d549923" exitCode=0 Feb 19 09:20:25 crc kubenswrapper[4675]: I0219 09:20:25.113154 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" event={"ID":"12901b5c-afde-4017-86ba-363d50c78a9a","Type":"ContainerDied","Data":"02a77cc089ac5389bd7e96e4434f2e59012b0743a1fa26d5709fd42f1d549923"} Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.593629 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.755137 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-ssh-key-openstack-edpm-ipam\") pod \"12901b5c-afde-4017-86ba-363d50c78a9a\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.755203 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/12901b5c-afde-4017-86ba-363d50c78a9a-kube-api-access-j6sht\") pod \"12901b5c-afde-4017-86ba-363d50c78a9a\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.755255 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-combined-ca-bundle\") pod \"12901b5c-afde-4017-86ba-363d50c78a9a\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.755364 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-inventory\") pod \"12901b5c-afde-4017-86ba-363d50c78a9a\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.755448 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-secret-0\") pod \"12901b5c-afde-4017-86ba-363d50c78a9a\" (UID: \"12901b5c-afde-4017-86ba-363d50c78a9a\") " Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.762244 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12901b5c-afde-4017-86ba-363d50c78a9a-kube-api-access-j6sht" (OuterVolumeSpecName: "kube-api-access-j6sht") pod "12901b5c-afde-4017-86ba-363d50c78a9a" (UID: "12901b5c-afde-4017-86ba-363d50c78a9a"). InnerVolumeSpecName "kube-api-access-j6sht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.767658 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "12901b5c-afde-4017-86ba-363d50c78a9a" (UID: "12901b5c-afde-4017-86ba-363d50c78a9a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.787094 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "12901b5c-afde-4017-86ba-363d50c78a9a" (UID: "12901b5c-afde-4017-86ba-363d50c78a9a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.787124 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "12901b5c-afde-4017-86ba-363d50c78a9a" (UID: "12901b5c-afde-4017-86ba-363d50c78a9a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.791326 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-inventory" (OuterVolumeSpecName: "inventory") pod "12901b5c-afde-4017-86ba-363d50c78a9a" (UID: "12901b5c-afde-4017-86ba-363d50c78a9a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.858703 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.858748 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/12901b5c-afde-4017-86ba-363d50c78a9a-kube-api-access-j6sht\") on node \"crc\" DevicePath \"\"" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.858758 4675 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.858768 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:20:26 crc kubenswrapper[4675]: I0219 09:20:26.858777 4675 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12901b5c-afde-4017-86ba-363d50c78a9a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.131169 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" event={"ID":"12901b5c-afde-4017-86ba-363d50c78a9a","Type":"ContainerDied","Data":"cce3cfa5f310c253fb711554a2a74477ea7633e1a0e6fcd1430108ea48d23b67"} Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.131228 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cce3cfa5f310c253fb711554a2a74477ea7633e1a0e6fcd1430108ea48d23b67" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.131281 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pknwk" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.237981 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h"] Feb 19 09:20:27 crc kubenswrapper[4675]: E0219 09:20:27.238387 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12901b5c-afde-4017-86ba-363d50c78a9a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.238410 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="12901b5c-afde-4017-86ba-363d50c78a9a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.238620 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="12901b5c-afde-4017-86ba-363d50c78a9a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.241405 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.245031 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.245304 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.245550 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.245802 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.245962 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.246117 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.250310 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.255535 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h"] Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.372112 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.372212 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.372404 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.372552 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.372752 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.372841 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkbzv\" (UniqueName: \"kubernetes.io/projected/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-kube-api-access-mkbzv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.373048 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.373150 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.373179 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.373347 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.373414 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475459 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475534 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475581 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475613 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475666 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475709 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475734 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkbzv\" (UniqueName: \"kubernetes.io/projected/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-kube-api-access-mkbzv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475772 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475809 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475825 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.475860 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.477004 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.480473 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.481214 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.481275 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.481844 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.482040 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.482110 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.482577 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.491383 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.492783 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.493844 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkbzv\" (UniqueName: \"kubernetes.io/projected/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-kube-api-access-mkbzv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-r2x4h\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:27 crc kubenswrapper[4675]: I0219 09:20:27.559271 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:20:28 crc kubenswrapper[4675]: I0219 09:20:28.104961 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h"] Feb 19 09:20:28 crc kubenswrapper[4675]: I0219 09:20:28.140468 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" event={"ID":"9cee21fc-60b4-413c-b6b4-e75e813a8e9d","Type":"ContainerStarted","Data":"a1d7e58fafdb7192d036a8790427dcfc8376b5f746493db8241f1908c919d05c"} Feb 19 09:20:29 crc kubenswrapper[4675]: I0219 09:20:29.158367 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" event={"ID":"9cee21fc-60b4-413c-b6b4-e75e813a8e9d","Type":"ContainerStarted","Data":"031ecd50d7a653ba929d8fc3e22ab277809e635fbe23af589c0c7bbcb76f72ba"} Feb 19 09:20:29 crc kubenswrapper[4675]: I0219 09:20:29.183102 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" podStartSLOduration=1.811195449 podStartE2EDuration="2.183077426s" podCreationTimestamp="2026-02-19 09:20:27 +0000 UTC" firstStartedPulling="2026-02-19 09:20:28.108693756 +0000 UTC m=+2209.735784024" lastFinishedPulling="2026-02-19 09:20:28.480575733 +0000 UTC m=+2210.107666001" observedRunningTime="2026-02-19 09:20:29.174530418 +0000 UTC m=+2210.801620696" watchObservedRunningTime="2026-02-19 09:20:29.183077426 +0000 UTC m=+2210.810167694" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.668916 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kd7c6"] Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.672534 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.684043 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kd7c6"] Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.717945 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8tgj\" (UniqueName: \"kubernetes.io/projected/ff0f631d-e86d-45b5-a75f-d6cf10835300-kube-api-access-n8tgj\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.718320 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-catalog-content\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.718443 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-utilities\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.820362 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8tgj\" (UniqueName: \"kubernetes.io/projected/ff0f631d-e86d-45b5-a75f-d6cf10835300-kube-api-access-n8tgj\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.820500 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-catalog-content\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.820527 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-utilities\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.821461 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-utilities\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.821592 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-catalog-content\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:43 crc kubenswrapper[4675]: I0219 09:20:43.844055 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8tgj\" (UniqueName: \"kubernetes.io/projected/ff0f631d-e86d-45b5-a75f-d6cf10835300-kube-api-access-n8tgj\") pod \"redhat-operators-kd7c6\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:44 crc kubenswrapper[4675]: I0219 09:20:44.011497 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:44 crc kubenswrapper[4675]: I0219 09:20:44.522288 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kd7c6"] Feb 19 09:20:44 crc kubenswrapper[4675]: W0219 09:20:44.522864 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff0f631d_e86d_45b5_a75f_d6cf10835300.slice/crio-f2ce77bfc9835e1f1b5f70aa47880930a8cfceb6bdb3fedc378ee5cc3ff062d7 WatchSource:0}: Error finding container f2ce77bfc9835e1f1b5f70aa47880930a8cfceb6bdb3fedc378ee5cc3ff062d7: Status 404 returned error can't find the container with id f2ce77bfc9835e1f1b5f70aa47880930a8cfceb6bdb3fedc378ee5cc3ff062d7 Feb 19 09:20:45 crc kubenswrapper[4675]: I0219 09:20:45.308717 4675 generic.go:334] "Generic (PLEG): container finished" podID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerID="1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1" exitCode=0 Feb 19 09:20:45 crc kubenswrapper[4675]: I0219 09:20:45.308826 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd7c6" event={"ID":"ff0f631d-e86d-45b5-a75f-d6cf10835300","Type":"ContainerDied","Data":"1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1"} Feb 19 09:20:45 crc kubenswrapper[4675]: I0219 09:20:45.309244 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd7c6" event={"ID":"ff0f631d-e86d-45b5-a75f-d6cf10835300","Type":"ContainerStarted","Data":"f2ce77bfc9835e1f1b5f70aa47880930a8cfceb6bdb3fedc378ee5cc3ff062d7"} Feb 19 09:20:47 crc kubenswrapper[4675]: I0219 09:20:47.352490 4675 generic.go:334] "Generic (PLEG): container finished" podID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerID="dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f" exitCode=0 Feb 19 09:20:47 crc kubenswrapper[4675]: I0219 09:20:47.352622 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd7c6" event={"ID":"ff0f631d-e86d-45b5-a75f-d6cf10835300","Type":"ContainerDied","Data":"dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f"} Feb 19 09:20:52 crc kubenswrapper[4675]: I0219 09:20:52.399110 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd7c6" event={"ID":"ff0f631d-e86d-45b5-a75f-d6cf10835300","Type":"ContainerStarted","Data":"233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8"} Feb 19 09:20:52 crc kubenswrapper[4675]: I0219 09:20:52.416439 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kd7c6" podStartSLOduration=2.95196991 podStartE2EDuration="9.416421104s" podCreationTimestamp="2026-02-19 09:20:43 +0000 UTC" firstStartedPulling="2026-02-19 09:20:45.310558776 +0000 UTC m=+2226.937649044" lastFinishedPulling="2026-02-19 09:20:51.77500997 +0000 UTC m=+2233.402100238" observedRunningTime="2026-02-19 09:20:52.414223316 +0000 UTC m=+2234.041313584" watchObservedRunningTime="2026-02-19 09:20:52.416421104 +0000 UTC m=+2234.043511372" Feb 19 09:20:54 crc kubenswrapper[4675]: I0219 09:20:54.013151 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:54 crc kubenswrapper[4675]: I0219 09:20:54.013703 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:20:55 crc kubenswrapper[4675]: I0219 09:20:55.064717 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kd7c6" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="registry-server" probeResult="failure" output=< Feb 19 09:20:55 crc kubenswrapper[4675]: timeout: failed to connect service ":50051" within 1s Feb 19 09:20:55 crc kubenswrapper[4675]: > Feb 19 09:21:04 crc kubenswrapper[4675]: I0219 09:21:04.060596 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:21:04 crc kubenswrapper[4675]: I0219 09:21:04.120465 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:21:04 crc kubenswrapper[4675]: I0219 09:21:04.305208 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kd7c6"] Feb 19 09:21:05 crc kubenswrapper[4675]: I0219 09:21:05.514797 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kd7c6" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="registry-server" containerID="cri-o://233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8" gracePeriod=2 Feb 19 09:21:05 crc kubenswrapper[4675]: I0219 09:21:05.988570 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.099120 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-catalog-content\") pod \"ff0f631d-e86d-45b5-a75f-d6cf10835300\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.099187 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8tgj\" (UniqueName: \"kubernetes.io/projected/ff0f631d-e86d-45b5-a75f-d6cf10835300-kube-api-access-n8tgj\") pod \"ff0f631d-e86d-45b5-a75f-d6cf10835300\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.099212 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-utilities\") pod \"ff0f631d-e86d-45b5-a75f-d6cf10835300\" (UID: \"ff0f631d-e86d-45b5-a75f-d6cf10835300\") " Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.100479 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-utilities" (OuterVolumeSpecName: "utilities") pod "ff0f631d-e86d-45b5-a75f-d6cf10835300" (UID: "ff0f631d-e86d-45b5-a75f-d6cf10835300"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.106364 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0f631d-e86d-45b5-a75f-d6cf10835300-kube-api-access-n8tgj" (OuterVolumeSpecName: "kube-api-access-n8tgj") pod "ff0f631d-e86d-45b5-a75f-d6cf10835300" (UID: "ff0f631d-e86d-45b5-a75f-d6cf10835300"). InnerVolumeSpecName "kube-api-access-n8tgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.202089 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8tgj\" (UniqueName: \"kubernetes.io/projected/ff0f631d-e86d-45b5-a75f-d6cf10835300-kube-api-access-n8tgj\") on node \"crc\" DevicePath \"\"" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.202485 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.219149 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff0f631d-e86d-45b5-a75f-d6cf10835300" (UID: "ff0f631d-e86d-45b5-a75f-d6cf10835300"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.304902 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0f631d-e86d-45b5-a75f-d6cf10835300-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.525152 4675 generic.go:334] "Generic (PLEG): container finished" podID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerID="233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8" exitCode=0 Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.525205 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd7c6" event={"ID":"ff0f631d-e86d-45b5-a75f-d6cf10835300","Type":"ContainerDied","Data":"233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8"} Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.525240 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd7c6" event={"ID":"ff0f631d-e86d-45b5-a75f-d6cf10835300","Type":"ContainerDied","Data":"f2ce77bfc9835e1f1b5f70aa47880930a8cfceb6bdb3fedc378ee5cc3ff062d7"} Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.525251 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd7c6" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.525262 4675 scope.go:117] "RemoveContainer" containerID="233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.556712 4675 scope.go:117] "RemoveContainer" containerID="dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.561527 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kd7c6"] Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.568874 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kd7c6"] Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.586072 4675 scope.go:117] "RemoveContainer" containerID="1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.623741 4675 scope.go:117] "RemoveContainer" containerID="233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8" Feb 19 09:21:06 crc kubenswrapper[4675]: E0219 09:21:06.624318 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8\": container with ID starting with 233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8 not found: ID does not exist" containerID="233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.624369 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8"} err="failed to get container status \"233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8\": rpc error: code = NotFound desc = could not find container \"233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8\": container with ID starting with 233d9ede972837854128c1ab6870f99575e0b3aa55832a40c2c175f34abd37c8 not found: ID does not exist" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.624400 4675 scope.go:117] "RemoveContainer" containerID="dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f" Feb 19 09:21:06 crc kubenswrapper[4675]: E0219 09:21:06.624780 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f\": container with ID starting with dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f not found: ID does not exist" containerID="dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.624815 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f"} err="failed to get container status \"dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f\": rpc error: code = NotFound desc = could not find container \"dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f\": container with ID starting with dfc91edcac3f1355ca301529e19b6afa40fe568b6f2b8a7e705006edf448d59f not found: ID does not exist" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.624835 4675 scope.go:117] "RemoveContainer" containerID="1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1" Feb 19 09:21:06 crc kubenswrapper[4675]: E0219 09:21:06.625166 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1\": container with ID starting with 1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1 not found: ID does not exist" containerID="1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1" Feb 19 09:21:06 crc kubenswrapper[4675]: I0219 09:21:06.625200 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1"} err="failed to get container status \"1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1\": rpc error: code = NotFound desc = could not find container \"1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1\": container with ID starting with 1b39aaeb7fefd6a4641cd5a02d3d329f62442d1acc2a5e0aaee03d7855303da1 not found: ID does not exist" Feb 19 09:21:07 crc kubenswrapper[4675]: I0219 09:21:07.114796 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" path="/var/lib/kubelet/pods/ff0f631d-e86d-45b5-a75f-d6cf10835300/volumes" Feb 19 09:22:35 crc kubenswrapper[4675]: I0219 09:22:35.282574 4675 generic.go:334] "Generic (PLEG): container finished" podID="9cee21fc-60b4-413c-b6b4-e75e813a8e9d" containerID="031ecd50d7a653ba929d8fc3e22ab277809e635fbe23af589c0c7bbcb76f72ba" exitCode=0 Feb 19 09:22:35 crc kubenswrapper[4675]: I0219 09:22:35.282677 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" event={"ID":"9cee21fc-60b4-413c-b6b4-e75e813a8e9d","Type":"ContainerDied","Data":"031ecd50d7a653ba929d8fc3e22ab277809e635fbe23af589c0c7bbcb76f72ba"} Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.706308 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.898164 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkbzv\" (UniqueName: \"kubernetes.io/projected/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-kube-api-access-mkbzv\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.898617 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-1\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.898675 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-extra-config-0\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.898734 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-0\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.898890 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-ssh-key-openstack-edpm-ipam\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.898975 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-combined-ca-bundle\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.899006 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-3\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.899055 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-2\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.899084 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-0\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.899108 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-inventory\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.899173 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-1\") pod \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\" (UID: \"9cee21fc-60b4-413c-b6b4-e75e813a8e9d\") " Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.908860 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.914733 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-kube-api-access-mkbzv" (OuterVolumeSpecName: "kube-api-access-mkbzv") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "kube-api-access-mkbzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.929034 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.933405 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.934410 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.934956 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.935624 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-inventory" (OuterVolumeSpecName: "inventory") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.936583 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.943544 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.951865 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:36 crc kubenswrapper[4675]: I0219 09:22:36.953690 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9cee21fc-60b4-413c-b6b4-e75e813a8e9d" (UID: "9cee21fc-60b4-413c-b6b4-e75e813a8e9d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002078 4675 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002123 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002137 4675 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002150 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkbzv\" (UniqueName: \"kubernetes.io/projected/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-kube-api-access-mkbzv\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002162 4675 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002174 4675 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002190 4675 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002204 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002216 4675 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002227 4675 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.002239 4675 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/9cee21fc-60b4-413c-b6b4-e75e813a8e9d-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.303867 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" event={"ID":"9cee21fc-60b4-413c-b6b4-e75e813a8e9d","Type":"ContainerDied","Data":"a1d7e58fafdb7192d036a8790427dcfc8376b5f746493db8241f1908c919d05c"} Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.304225 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1d7e58fafdb7192d036a8790427dcfc8376b5f746493db8241f1908c919d05c" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.304198 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-r2x4h" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.417606 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl"] Feb 19 09:22:37 crc kubenswrapper[4675]: E0219 09:22:37.418200 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="extract-content" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.418225 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="extract-content" Feb 19 09:22:37 crc kubenswrapper[4675]: E0219 09:22:37.418246 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cee21fc-60b4-413c-b6b4-e75e813a8e9d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.418255 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cee21fc-60b4-413c-b6b4-e75e813a8e9d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 19 09:22:37 crc kubenswrapper[4675]: E0219 09:22:37.418263 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="extract-utilities" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.418272 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="extract-utilities" Feb 19 09:22:37 crc kubenswrapper[4675]: E0219 09:22:37.418286 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="registry-server" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.418293 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="registry-server" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.418516 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0f631d-e86d-45b5-a75f-d6cf10835300" containerName="registry-server" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.418545 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cee21fc-60b4-413c-b6b4-e75e813a8e9d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.419595 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.427827 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.427870 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mr8b" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.427969 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.427795 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl"] Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.428150 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.428227 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.516949 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.517073 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cmvm\" (UniqueName: \"kubernetes.io/projected/c1b6f896-16c6-40af-ae5f-4392213dacd3-kube-api-access-6cmvm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.517130 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.517211 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.517280 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.517427 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.517556 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.619613 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.619756 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.619827 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.619855 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cmvm\" (UniqueName: \"kubernetes.io/projected/c1b6f896-16c6-40af-ae5f-4392213dacd3-kube-api-access-6cmvm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.619886 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.619923 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.619956 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.624343 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.627238 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.631346 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.632311 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.632595 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.632731 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.636956 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cmvm\" (UniqueName: \"kubernetes.io/projected/c1b6f896-16c6-40af-ae5f-4392213dacd3-kube-api-access-6cmvm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:37 crc kubenswrapper[4675]: I0219 09:22:37.738464 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:22:38 crc kubenswrapper[4675]: I0219 09:22:38.249844 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl"] Feb 19 09:22:38 crc kubenswrapper[4675]: I0219 09:22:38.255674 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:22:38 crc kubenswrapper[4675]: I0219 09:22:38.314148 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" event={"ID":"c1b6f896-16c6-40af-ae5f-4392213dacd3","Type":"ContainerStarted","Data":"a7974bc07bea43181b033b36645d225d884dc72527388ba00008416bf625425d"} Feb 19 09:22:39 crc kubenswrapper[4675]: I0219 09:22:39.335392 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" event={"ID":"c1b6f896-16c6-40af-ae5f-4392213dacd3","Type":"ContainerStarted","Data":"ebdc82f8868c3769d3f8a7dfc315295bd1e3545de5b6d61473bbc4fb044e7973"} Feb 19 09:22:39 crc kubenswrapper[4675]: I0219 09:22:39.355310 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" podStartSLOduration=1.93334495 podStartE2EDuration="2.35529274s" podCreationTimestamp="2026-02-19 09:22:37 +0000 UTC" firstStartedPulling="2026-02-19 09:22:38.255468496 +0000 UTC m=+2339.882558764" lastFinishedPulling="2026-02-19 09:22:38.677416266 +0000 UTC m=+2340.304506554" observedRunningTime="2026-02-19 09:22:39.352270659 +0000 UTC m=+2340.979360927" watchObservedRunningTime="2026-02-19 09:22:39.35529274 +0000 UTC m=+2340.982383008" Feb 19 09:22:41 crc kubenswrapper[4675]: I0219 09:22:41.744669 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:22:41 crc kubenswrapper[4675]: I0219 09:22:41.745264 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:23:11 crc kubenswrapper[4675]: I0219 09:23:11.743406 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:23:11 crc kubenswrapper[4675]: I0219 09:23:11.744005 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.743539 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.744354 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.744447 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.745661 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.745729 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" gracePeriod=600 Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.896812 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" exitCode=0 Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.897273 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b"} Feb 19 09:23:41 crc kubenswrapper[4675]: I0219 09:23:41.897450 4675 scope.go:117] "RemoveContainer" containerID="e7b56eaf6b398d1eb891e81f3edb6374db4978daeafa1c5ebab65fecd4bcb293" Feb 19 09:23:41 crc kubenswrapper[4675]: E0219 09:23:41.899718 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:23:42 crc kubenswrapper[4675]: I0219 09:23:42.920797 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:23:42 crc kubenswrapper[4675]: E0219 09:23:42.921599 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:23:57 crc kubenswrapper[4675]: I0219 09:23:57.103786 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:23:57 crc kubenswrapper[4675]: E0219 09:23:57.104829 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:24:08 crc kubenswrapper[4675]: I0219 09:24:08.104251 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:24:08 crc kubenswrapper[4675]: E0219 09:24:08.105337 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:24:21 crc kubenswrapper[4675]: I0219 09:24:21.104242 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:24:21 crc kubenswrapper[4675]: E0219 09:24:21.105265 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:24:36 crc kubenswrapper[4675]: I0219 09:24:36.103328 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:24:36 crc kubenswrapper[4675]: E0219 09:24:36.104196 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:24:44 crc kubenswrapper[4675]: I0219 09:24:44.403053 4675 generic.go:334] "Generic (PLEG): container finished" podID="c1b6f896-16c6-40af-ae5f-4392213dacd3" containerID="ebdc82f8868c3769d3f8a7dfc315295bd1e3545de5b6d61473bbc4fb044e7973" exitCode=0 Feb 19 09:24:44 crc kubenswrapper[4675]: I0219 09:24:44.403140 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" event={"ID":"c1b6f896-16c6-40af-ae5f-4392213dacd3","Type":"ContainerDied","Data":"ebdc82f8868c3769d3f8a7dfc315295bd1e3545de5b6d61473bbc4fb044e7973"} Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.801253 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.901152 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ssh-key-openstack-edpm-ipam\") pod \"c1b6f896-16c6-40af-ae5f-4392213dacd3\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.901276 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-inventory\") pod \"c1b6f896-16c6-40af-ae5f-4392213dacd3\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.901320 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-0\") pod \"c1b6f896-16c6-40af-ae5f-4392213dacd3\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.901438 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-2\") pod \"c1b6f896-16c6-40af-ae5f-4392213dacd3\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.901515 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cmvm\" (UniqueName: \"kubernetes.io/projected/c1b6f896-16c6-40af-ae5f-4392213dacd3-kube-api-access-6cmvm\") pod \"c1b6f896-16c6-40af-ae5f-4392213dacd3\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.901571 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-1\") pod \"c1b6f896-16c6-40af-ae5f-4392213dacd3\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.901631 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-telemetry-combined-ca-bundle\") pod \"c1b6f896-16c6-40af-ae5f-4392213dacd3\" (UID: \"c1b6f896-16c6-40af-ae5f-4392213dacd3\") " Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.910927 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c1b6f896-16c6-40af-ae5f-4392213dacd3" (UID: "c1b6f896-16c6-40af-ae5f-4392213dacd3"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.957936 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b6f896-16c6-40af-ae5f-4392213dacd3-kube-api-access-6cmvm" (OuterVolumeSpecName: "kube-api-access-6cmvm") pod "c1b6f896-16c6-40af-ae5f-4392213dacd3" (UID: "c1b6f896-16c6-40af-ae5f-4392213dacd3"). InnerVolumeSpecName "kube-api-access-6cmvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.971886 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-inventory" (OuterVolumeSpecName: "inventory") pod "c1b6f896-16c6-40af-ae5f-4392213dacd3" (UID: "c1b6f896-16c6-40af-ae5f-4392213dacd3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:24:45 crc kubenswrapper[4675]: I0219 09:24:45.981803 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c1b6f896-16c6-40af-ae5f-4392213dacd3" (UID: "c1b6f896-16c6-40af-ae5f-4392213dacd3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.008914 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "c1b6f896-16c6-40af-ae5f-4392213dacd3" (UID: "c1b6f896-16c6-40af-ae5f-4392213dacd3"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.009046 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "c1b6f896-16c6-40af-ae5f-4392213dacd3" (UID: "c1b6f896-16c6-40af-ae5f-4392213dacd3"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.011216 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cmvm\" (UniqueName: \"kubernetes.io/projected/c1b6f896-16c6-40af-ae5f-4392213dacd3-kube-api-access-6cmvm\") on node \"crc\" DevicePath \"\"" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.011270 4675 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.011282 4675 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.011292 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.011302 4675 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-inventory\") on node \"crc\" DevicePath \"\"" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.011311 4675 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.077733 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "c1b6f896-16c6-40af-ae5f-4392213dacd3" (UID: "c1b6f896-16c6-40af-ae5f-4392213dacd3"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.113431 4675 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c1b6f896-16c6-40af-ae5f-4392213dacd3-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.423010 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" event={"ID":"c1b6f896-16c6-40af-ae5f-4392213dacd3","Type":"ContainerDied","Data":"a7974bc07bea43181b033b36645d225d884dc72527388ba00008416bf625425d"} Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.423555 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7974bc07bea43181b033b36645d225d884dc72527388ba00008416bf625425d" Feb 19 09:24:46 crc kubenswrapper[4675]: I0219 09:24:46.423071 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl" Feb 19 09:24:47 crc kubenswrapper[4675]: I0219 09:24:47.103534 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:24:47 crc kubenswrapper[4675]: E0219 09:24:47.103914 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:25:01 crc kubenswrapper[4675]: I0219 09:25:01.104186 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:25:01 crc kubenswrapper[4675]: E0219 09:25:01.105618 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:25:12 crc kubenswrapper[4675]: I0219 09:25:12.103867 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:25:12 crc kubenswrapper[4675]: E0219 09:25:12.105215 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.902624 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hw7q2"] Feb 19 09:25:18 crc kubenswrapper[4675]: E0219 09:25:18.903902 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b6f896-16c6-40af-ae5f-4392213dacd3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.903928 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b6f896-16c6-40af-ae5f-4392213dacd3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.904089 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b6f896-16c6-40af-ae5f-4392213dacd3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.905426 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.928907 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw7q2"] Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.962430 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-utilities\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.962509 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwgjc\" (UniqueName: \"kubernetes.io/projected/4d03776c-adef-4f5f-af13-91290e06a2d3-kube-api-access-fwgjc\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:18 crc kubenswrapper[4675]: I0219 09:25:18.962545 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-catalog-content\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.064486 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-utilities\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.064560 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwgjc\" (UniqueName: \"kubernetes.io/projected/4d03776c-adef-4f5f-af13-91290e06a2d3-kube-api-access-fwgjc\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.064595 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-catalog-content\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.065063 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-utilities\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.065146 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-catalog-content\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.087748 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwgjc\" (UniqueName: \"kubernetes.io/projected/4d03776c-adef-4f5f-af13-91290e06a2d3-kube-api-access-fwgjc\") pod \"redhat-marketplace-hw7q2\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.232325 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.799987 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw7q2"] Feb 19 09:25:19 crc kubenswrapper[4675]: I0219 09:25:19.929510 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw7q2" event={"ID":"4d03776c-adef-4f5f-af13-91290e06a2d3","Type":"ContainerStarted","Data":"7646482e6cf178f9cefe403aa00f9d3e40487b548e265a8ad5f170abfbfdb17e"} Feb 19 09:25:20 crc kubenswrapper[4675]: I0219 09:25:20.940143 4675 generic.go:334] "Generic (PLEG): container finished" podID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerID="0478cea111e218b631f4414afab89ba3237db7ca7ac3725a02b88e41a62c93e6" exitCode=0 Feb 19 09:25:20 crc kubenswrapper[4675]: I0219 09:25:20.940237 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw7q2" event={"ID":"4d03776c-adef-4f5f-af13-91290e06a2d3","Type":"ContainerDied","Data":"0478cea111e218b631f4414afab89ba3237db7ca7ac3725a02b88e41a62c93e6"} Feb 19 09:25:21 crc kubenswrapper[4675]: I0219 09:25:21.950762 4675 generic.go:334] "Generic (PLEG): container finished" podID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerID="ceb62e8d3dcce61b09d307788deaf22476174c816a814f937e92d9e079a23732" exitCode=0 Feb 19 09:25:21 crc kubenswrapper[4675]: I0219 09:25:21.950888 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw7q2" event={"ID":"4d03776c-adef-4f5f-af13-91290e06a2d3","Type":"ContainerDied","Data":"ceb62e8d3dcce61b09d307788deaf22476174c816a814f937e92d9e079a23732"} Feb 19 09:25:22 crc kubenswrapper[4675]: I0219 09:25:22.978532 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw7q2" event={"ID":"4d03776c-adef-4f5f-af13-91290e06a2d3","Type":"ContainerStarted","Data":"fb2667ed30062eba8231fdd98bbee74de48acb72f403b33aa2da3da51ff35e17"} Feb 19 09:25:23 crc kubenswrapper[4675]: I0219 09:25:23.015011 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hw7q2" podStartSLOduration=3.550582902 podStartE2EDuration="5.014990099s" podCreationTimestamp="2026-02-19 09:25:18 +0000 UTC" firstStartedPulling="2026-02-19 09:25:20.944493191 +0000 UTC m=+2502.571583469" lastFinishedPulling="2026-02-19 09:25:22.408900398 +0000 UTC m=+2504.035990666" observedRunningTime="2026-02-19 09:25:23.004257652 +0000 UTC m=+2504.631347930" watchObservedRunningTime="2026-02-19 09:25:23.014990099 +0000 UTC m=+2504.642080367" Feb 19 09:25:26 crc kubenswrapper[4675]: I0219 09:25:26.104134 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:25:26 crc kubenswrapper[4675]: E0219 09:25:26.104829 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.559168 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.562165 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.565065 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.565128 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.565372 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-h67kp" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.565526 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.586813 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.677526 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.677663 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.677737 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.677898 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.678150 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-config-data\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.678342 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.678388 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scl6w\" (UniqueName: \"kubernetes.io/projected/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-kube-api-access-scl6w\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.678471 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.678570 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.780845 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.780935 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.780961 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.781006 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.781034 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.781109 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-config-data\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.781158 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.781234 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scl6w\" (UniqueName: \"kubernetes.io/projected/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-kube-api-access-scl6w\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.781269 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.782548 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.782720 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.782863 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.782933 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-config-data\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.782963 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.788204 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.788231 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.790835 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.798841 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scl6w\" (UniqueName: \"kubernetes.io/projected/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-kube-api-access-scl6w\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.811418 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " pod="openstack/tempest-tests-tempest" Feb 19 09:25:28 crc kubenswrapper[4675]: I0219 09:25:28.891988 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 19 09:25:29 crc kubenswrapper[4675]: I0219 09:25:29.233004 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:29 crc kubenswrapper[4675]: I0219 09:25:29.233453 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:29 crc kubenswrapper[4675]: I0219 09:25:29.283415 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:29 crc kubenswrapper[4675]: I0219 09:25:29.353084 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 19 09:25:30 crc kubenswrapper[4675]: I0219 09:25:30.040618 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5006db86-ae9d-4dbc-871f-4bc18f3a43f6","Type":"ContainerStarted","Data":"4e1d7d6533475cf6e7c0cf274177ea15c311b225b3eb205fcdc0c64a072008f2"} Feb 19 09:25:30 crc kubenswrapper[4675]: I0219 09:25:30.097455 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:30 crc kubenswrapper[4675]: I0219 09:25:30.159056 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw7q2"] Feb 19 09:25:32 crc kubenswrapper[4675]: I0219 09:25:32.070230 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hw7q2" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="registry-server" containerID="cri-o://fb2667ed30062eba8231fdd98bbee74de48acb72f403b33aa2da3da51ff35e17" gracePeriod=2 Feb 19 09:25:33 crc kubenswrapper[4675]: I0219 09:25:33.085986 4675 generic.go:334] "Generic (PLEG): container finished" podID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerID="fb2667ed30062eba8231fdd98bbee74de48acb72f403b33aa2da3da51ff35e17" exitCode=0 Feb 19 09:25:33 crc kubenswrapper[4675]: I0219 09:25:33.086059 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw7q2" event={"ID":"4d03776c-adef-4f5f-af13-91290e06a2d3","Type":"ContainerDied","Data":"fb2667ed30062eba8231fdd98bbee74de48acb72f403b33aa2da3da51ff35e17"} Feb 19 09:25:34 crc kubenswrapper[4675]: I0219 09:25:34.875425 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.010903 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwgjc\" (UniqueName: \"kubernetes.io/projected/4d03776c-adef-4f5f-af13-91290e06a2d3-kube-api-access-fwgjc\") pod \"4d03776c-adef-4f5f-af13-91290e06a2d3\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.011163 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-catalog-content\") pod \"4d03776c-adef-4f5f-af13-91290e06a2d3\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.011248 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-utilities\") pod \"4d03776c-adef-4f5f-af13-91290e06a2d3\" (UID: \"4d03776c-adef-4f5f-af13-91290e06a2d3\") " Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.012151 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-utilities" (OuterVolumeSpecName: "utilities") pod "4d03776c-adef-4f5f-af13-91290e06a2d3" (UID: "4d03776c-adef-4f5f-af13-91290e06a2d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.019430 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d03776c-adef-4f5f-af13-91290e06a2d3-kube-api-access-fwgjc" (OuterVolumeSpecName: "kube-api-access-fwgjc") pod "4d03776c-adef-4f5f-af13-91290e06a2d3" (UID: "4d03776c-adef-4f5f-af13-91290e06a2d3"). InnerVolumeSpecName "kube-api-access-fwgjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.036417 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d03776c-adef-4f5f-af13-91290e06a2d3" (UID: "4d03776c-adef-4f5f-af13-91290e06a2d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.109282 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hw7q2" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.113660 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.113689 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d03776c-adef-4f5f-af13-91290e06a2d3-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.113699 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwgjc\" (UniqueName: \"kubernetes.io/projected/4d03776c-adef-4f5f-af13-91290e06a2d3-kube-api-access-fwgjc\") on node \"crc\" DevicePath \"\"" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.117781 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hw7q2" event={"ID":"4d03776c-adef-4f5f-af13-91290e06a2d3","Type":"ContainerDied","Data":"7646482e6cf178f9cefe403aa00f9d3e40487b548e265a8ad5f170abfbfdb17e"} Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.117836 4675 scope.go:117] "RemoveContainer" containerID="fb2667ed30062eba8231fdd98bbee74de48acb72f403b33aa2da3da51ff35e17" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.154961 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw7q2"] Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.163113 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hw7q2"] Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.165682 4675 scope.go:117] "RemoveContainer" containerID="ceb62e8d3dcce61b09d307788deaf22476174c816a814f937e92d9e079a23732" Feb 19 09:25:35 crc kubenswrapper[4675]: I0219 09:25:35.187229 4675 scope.go:117] "RemoveContainer" containerID="0478cea111e218b631f4414afab89ba3237db7ca7ac3725a02b88e41a62c93e6" Feb 19 09:25:37 crc kubenswrapper[4675]: I0219 09:25:37.114985 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" path="/var/lib/kubelet/pods/4d03776c-adef-4f5f-af13-91290e06a2d3/volumes" Feb 19 09:25:41 crc kubenswrapper[4675]: I0219 09:25:41.104295 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:25:41 crc kubenswrapper[4675]: E0219 09:25:41.105180 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.058069 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6vk46"] Feb 19 09:25:50 crc kubenswrapper[4675]: E0219 09:25:50.059342 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="registry-server" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.059355 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="registry-server" Feb 19 09:25:50 crc kubenswrapper[4675]: E0219 09:25:50.059388 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="extract-content" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.059394 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="extract-content" Feb 19 09:25:50 crc kubenswrapper[4675]: E0219 09:25:50.059411 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="extract-utilities" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.059417 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="extract-utilities" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.059616 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d03776c-adef-4f5f-af13-91290e06a2d3" containerName="registry-server" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.089623 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vk46"] Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.090167 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.194000 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-catalog-content\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.194523 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-utilities\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.194682 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhxqv\" (UniqueName: \"kubernetes.io/projected/7a07bcce-8d2f-4354-b830-0ecf95364345-kube-api-access-mhxqv\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.297696 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-utilities\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.298065 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-catalog-content\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.298185 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhxqv\" (UniqueName: \"kubernetes.io/projected/7a07bcce-8d2f-4354-b830-0ecf95364345-kube-api-access-mhxqv\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.298264 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-utilities\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.298939 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-catalog-content\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.322478 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhxqv\" (UniqueName: \"kubernetes.io/projected/7a07bcce-8d2f-4354-b830-0ecf95364345-kube-api-access-mhxqv\") pod \"certified-operators-6vk46\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:50 crc kubenswrapper[4675]: I0219 09:25:50.431181 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:25:56 crc kubenswrapper[4675]: I0219 09:25:56.102939 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:25:56 crc kubenswrapper[4675]: E0219 09:25:56.104105 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:25:58 crc kubenswrapper[4675]: E0219 09:25:58.689428 4675 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 19 09:25:58 crc kubenswrapper[4675]: E0219 09:25:58.690307 4675 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-scl6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(5006db86-ae9d-4dbc-871f-4bc18f3a43f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 19 09:25:58 crc kubenswrapper[4675]: E0219 09:25:58.691881 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="5006db86-ae9d-4dbc-871f-4bc18f3a43f6" Feb 19 09:25:59 crc kubenswrapper[4675]: I0219 09:25:59.166590 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vk46"] Feb 19 09:25:59 crc kubenswrapper[4675]: I0219 09:25:59.343331 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vk46" event={"ID":"7a07bcce-8d2f-4354-b830-0ecf95364345","Type":"ContainerStarted","Data":"1cc555a5e09b4b4765253299f890cbae26f2a514af2095dae69043778b0d18a0"} Feb 19 09:25:59 crc kubenswrapper[4675]: E0219 09:25:59.345233 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="5006db86-ae9d-4dbc-871f-4bc18f3a43f6" Feb 19 09:26:00 crc kubenswrapper[4675]: I0219 09:26:00.353028 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerID="9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9" exitCode=0 Feb 19 09:26:00 crc kubenswrapper[4675]: I0219 09:26:00.353047 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vk46" event={"ID":"7a07bcce-8d2f-4354-b830-0ecf95364345","Type":"ContainerDied","Data":"9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9"} Feb 19 09:26:01 crc kubenswrapper[4675]: I0219 09:26:01.364684 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vk46" event={"ID":"7a07bcce-8d2f-4354-b830-0ecf95364345","Type":"ContainerStarted","Data":"a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906"} Feb 19 09:26:02 crc kubenswrapper[4675]: I0219 09:26:02.374103 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerID="a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906" exitCode=0 Feb 19 09:26:02 crc kubenswrapper[4675]: I0219 09:26:02.374183 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vk46" event={"ID":"7a07bcce-8d2f-4354-b830-0ecf95364345","Type":"ContainerDied","Data":"a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906"} Feb 19 09:26:03 crc kubenswrapper[4675]: I0219 09:26:03.384749 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vk46" event={"ID":"7a07bcce-8d2f-4354-b830-0ecf95364345","Type":"ContainerStarted","Data":"e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7"} Feb 19 09:26:03 crc kubenswrapper[4675]: I0219 09:26:03.411178 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6vk46" podStartSLOduration=10.917331069 podStartE2EDuration="13.411158175s" podCreationTimestamp="2026-02-19 09:25:50 +0000 UTC" firstStartedPulling="2026-02-19 09:26:00.357440815 +0000 UTC m=+2541.984531073" lastFinishedPulling="2026-02-19 09:26:02.851267911 +0000 UTC m=+2544.478358179" observedRunningTime="2026-02-19 09:26:03.404096776 +0000 UTC m=+2545.031187044" watchObservedRunningTime="2026-02-19 09:26:03.411158175 +0000 UTC m=+2545.038248443" Feb 19 09:26:10 crc kubenswrapper[4675]: I0219 09:26:10.104869 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:26:10 crc kubenswrapper[4675]: E0219 09:26:10.105944 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:26:10 crc kubenswrapper[4675]: I0219 09:26:10.431884 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:26:10 crc kubenswrapper[4675]: I0219 09:26:10.431928 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:26:10 crc kubenswrapper[4675]: I0219 09:26:10.480890 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:26:10 crc kubenswrapper[4675]: I0219 09:26:10.525839 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:26:10 crc kubenswrapper[4675]: I0219 09:26:10.719055 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vk46"] Feb 19 09:26:12 crc kubenswrapper[4675]: I0219 09:26:12.462701 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6vk46" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="registry-server" containerID="cri-o://e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7" gracePeriod=2 Feb 19 09:26:12 crc kubenswrapper[4675]: I0219 09:26:12.920696 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.074026 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhxqv\" (UniqueName: \"kubernetes.io/projected/7a07bcce-8d2f-4354-b830-0ecf95364345-kube-api-access-mhxqv\") pod \"7a07bcce-8d2f-4354-b830-0ecf95364345\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.074308 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-catalog-content\") pod \"7a07bcce-8d2f-4354-b830-0ecf95364345\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.074418 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-utilities\") pod \"7a07bcce-8d2f-4354-b830-0ecf95364345\" (UID: \"7a07bcce-8d2f-4354-b830-0ecf95364345\") " Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.075287 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-utilities" (OuterVolumeSpecName: "utilities") pod "7a07bcce-8d2f-4354-b830-0ecf95364345" (UID: "7a07bcce-8d2f-4354-b830-0ecf95364345"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.082175 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a07bcce-8d2f-4354-b830-0ecf95364345-kube-api-access-mhxqv" (OuterVolumeSpecName: "kube-api-access-mhxqv") pod "7a07bcce-8d2f-4354-b830-0ecf95364345" (UID: "7a07bcce-8d2f-4354-b830-0ecf95364345"). InnerVolumeSpecName "kube-api-access-mhxqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.143165 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a07bcce-8d2f-4354-b830-0ecf95364345" (UID: "7a07bcce-8d2f-4354-b830-0ecf95364345"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.177485 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.177541 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhxqv\" (UniqueName: \"kubernetes.io/projected/7a07bcce-8d2f-4354-b830-0ecf95364345-kube-api-access-mhxqv\") on node \"crc\" DevicePath \"\"" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.177555 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a07bcce-8d2f-4354-b830-0ecf95364345-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.472386 4675 generic.go:334] "Generic (PLEG): container finished" podID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerID="e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7" exitCode=0 Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.472609 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vk46" event={"ID":"7a07bcce-8d2f-4354-b830-0ecf95364345","Type":"ContainerDied","Data":"e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7"} Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.473133 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vk46" event={"ID":"7a07bcce-8d2f-4354-b830-0ecf95364345","Type":"ContainerDied","Data":"1cc555a5e09b4b4765253299f890cbae26f2a514af2095dae69043778b0d18a0"} Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.473174 4675 scope.go:117] "RemoveContainer" containerID="e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.472890 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vk46" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.509464 4675 scope.go:117] "RemoveContainer" containerID="a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.514143 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vk46"] Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.525003 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6vk46"] Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.533446 4675 scope.go:117] "RemoveContainer" containerID="9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.580836 4675 scope.go:117] "RemoveContainer" containerID="e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7" Feb 19 09:26:13 crc kubenswrapper[4675]: E0219 09:26:13.581417 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7\": container with ID starting with e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7 not found: ID does not exist" containerID="e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.581461 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7"} err="failed to get container status \"e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7\": rpc error: code = NotFound desc = could not find container \"e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7\": container with ID starting with e0e279d5668e2d45d14cc2a35f6f5f646117be2d2387bc25bd56259b99803ae7 not found: ID does not exist" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.581487 4675 scope.go:117] "RemoveContainer" containerID="a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906" Feb 19 09:26:13 crc kubenswrapper[4675]: E0219 09:26:13.581837 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906\": container with ID starting with a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906 not found: ID does not exist" containerID="a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.581873 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906"} err="failed to get container status \"a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906\": rpc error: code = NotFound desc = could not find container \"a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906\": container with ID starting with a99c8be062c2026286c4adb7031e2b2ec87cb72441c44042a72569def1997906 not found: ID does not exist" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.581892 4675 scope.go:117] "RemoveContainer" containerID="9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9" Feb 19 09:26:13 crc kubenswrapper[4675]: E0219 09:26:13.582162 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9\": container with ID starting with 9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9 not found: ID does not exist" containerID="9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.582206 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9"} err="failed to get container status \"9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9\": rpc error: code = NotFound desc = could not find container \"9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9\": container with ID starting with 9b78016c6244a350e141de3a32b03c377a86db76994a8bf1a68fc5cd8b95d0a9 not found: ID does not exist" Feb 19 09:26:13 crc kubenswrapper[4675]: I0219 09:26:13.686780 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 19 09:26:15 crc kubenswrapper[4675]: I0219 09:26:15.116665 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" path="/var/lib/kubelet/pods/7a07bcce-8d2f-4354-b830-0ecf95364345/volumes" Feb 19 09:26:15 crc kubenswrapper[4675]: I0219 09:26:15.505392 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5006db86-ae9d-4dbc-871f-4bc18f3a43f6","Type":"ContainerStarted","Data":"b10817aa3eeed3cee7c7e92e420ba77914b0ec9c562c157f9cb0393cd4b7034b"} Feb 19 09:26:15 crc kubenswrapper[4675]: I0219 09:26:15.530978 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.206008789 podStartE2EDuration="48.53095811s" podCreationTimestamp="2026-02-19 09:25:27 +0000 UTC" firstStartedPulling="2026-02-19 09:25:29.359368685 +0000 UTC m=+2510.986458953" lastFinishedPulling="2026-02-19 09:26:13.684318006 +0000 UTC m=+2555.311408274" observedRunningTime="2026-02-19 09:26:15.522044641 +0000 UTC m=+2557.149134909" watchObservedRunningTime="2026-02-19 09:26:15.53095811 +0000 UTC m=+2557.158048378" Feb 19 09:26:24 crc kubenswrapper[4675]: I0219 09:26:24.103427 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:26:24 crc kubenswrapper[4675]: E0219 09:26:24.106311 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:26:36 crc kubenswrapper[4675]: I0219 09:26:36.103159 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:26:36 crc kubenswrapper[4675]: E0219 09:26:36.104494 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:26:49 crc kubenswrapper[4675]: I0219 09:26:49.110360 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:26:49 crc kubenswrapper[4675]: E0219 09:26:49.111552 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:27:01 crc kubenswrapper[4675]: I0219 09:27:01.104161 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:27:01 crc kubenswrapper[4675]: E0219 09:27:01.105361 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:27:13 crc kubenswrapper[4675]: I0219 09:27:13.103726 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:27:13 crc kubenswrapper[4675]: E0219 09:27:13.104819 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:27:26 crc kubenswrapper[4675]: I0219 09:27:26.102930 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:27:26 crc kubenswrapper[4675]: E0219 09:27:26.103875 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:27:38 crc kubenswrapper[4675]: I0219 09:27:38.103185 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:27:38 crc kubenswrapper[4675]: E0219 09:27:38.104110 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:27:49 crc kubenswrapper[4675]: I0219 09:27:49.110171 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:27:49 crc kubenswrapper[4675]: E0219 09:27:49.111688 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:28:04 crc kubenswrapper[4675]: I0219 09:28:04.103695 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:28:04 crc kubenswrapper[4675]: E0219 09:28:04.104711 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:28:16 crc kubenswrapper[4675]: I0219 09:28:16.104021 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:28:16 crc kubenswrapper[4675]: E0219 09:28:16.105079 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.707426 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bwfg6"] Feb 19 09:28:17 crc kubenswrapper[4675]: E0219 09:28:17.708451 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="extract-utilities" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.708467 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="extract-utilities" Feb 19 09:28:17 crc kubenswrapper[4675]: E0219 09:28:17.708487 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="registry-server" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.708495 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="registry-server" Feb 19 09:28:17 crc kubenswrapper[4675]: E0219 09:28:17.708530 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="extract-content" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.708537 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="extract-content" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.708808 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a07bcce-8d2f-4354-b830-0ecf95364345" containerName="registry-server" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.711580 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.748464 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwfg6"] Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.849308 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-utilities\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.849810 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-catalog-content\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.849875 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc9rs\" (UniqueName: \"kubernetes.io/projected/45a88869-9ddb-4751-b44c-b035f974d06f-kube-api-access-xc9rs\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.951938 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-catalog-content\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.952053 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc9rs\" (UniqueName: \"kubernetes.io/projected/45a88869-9ddb-4751-b44c-b035f974d06f-kube-api-access-xc9rs\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.952195 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-utilities\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.952555 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-catalog-content\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.952583 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-utilities\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:17 crc kubenswrapper[4675]: I0219 09:28:17.974347 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc9rs\" (UniqueName: \"kubernetes.io/projected/45a88869-9ddb-4751-b44c-b035f974d06f-kube-api-access-xc9rs\") pod \"community-operators-bwfg6\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:18 crc kubenswrapper[4675]: I0219 09:28:18.038846 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:18 crc kubenswrapper[4675]: I0219 09:28:18.570511 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwfg6"] Feb 19 09:28:18 crc kubenswrapper[4675]: I0219 09:28:18.643726 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwfg6" event={"ID":"45a88869-9ddb-4751-b44c-b035f974d06f","Type":"ContainerStarted","Data":"bb28e1b0d587eb0ac0841269c9c4ce04e598a0d079807dc3fb74f16413a48181"} Feb 19 09:28:19 crc kubenswrapper[4675]: I0219 09:28:19.652548 4675 generic.go:334] "Generic (PLEG): container finished" podID="45a88869-9ddb-4751-b44c-b035f974d06f" containerID="d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6" exitCode=0 Feb 19 09:28:19 crc kubenswrapper[4675]: I0219 09:28:19.654427 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwfg6" event={"ID":"45a88869-9ddb-4751-b44c-b035f974d06f","Type":"ContainerDied","Data":"d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6"} Feb 19 09:28:19 crc kubenswrapper[4675]: I0219 09:28:19.654459 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:28:20 crc kubenswrapper[4675]: I0219 09:28:20.663132 4675 generic.go:334] "Generic (PLEG): container finished" podID="45a88869-9ddb-4751-b44c-b035f974d06f" containerID="e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987" exitCode=0 Feb 19 09:28:20 crc kubenswrapper[4675]: I0219 09:28:20.663270 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwfg6" event={"ID":"45a88869-9ddb-4751-b44c-b035f974d06f","Type":"ContainerDied","Data":"e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987"} Feb 19 09:28:21 crc kubenswrapper[4675]: I0219 09:28:21.682521 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwfg6" event={"ID":"45a88869-9ddb-4751-b44c-b035f974d06f","Type":"ContainerStarted","Data":"fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4"} Feb 19 09:28:21 crc kubenswrapper[4675]: I0219 09:28:21.703106 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bwfg6" podStartSLOduration=3.297410691 podStartE2EDuration="4.703086724s" podCreationTimestamp="2026-02-19 09:28:17 +0000 UTC" firstStartedPulling="2026-02-19 09:28:19.654201126 +0000 UTC m=+2681.281291404" lastFinishedPulling="2026-02-19 09:28:21.059877169 +0000 UTC m=+2682.686967437" observedRunningTime="2026-02-19 09:28:21.702487878 +0000 UTC m=+2683.329578146" watchObservedRunningTime="2026-02-19 09:28:21.703086724 +0000 UTC m=+2683.330177002" Feb 19 09:28:27 crc kubenswrapper[4675]: I0219 09:28:27.104057 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:28:27 crc kubenswrapper[4675]: E0219 09:28:27.105020 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:28:28 crc kubenswrapper[4675]: I0219 09:28:28.039456 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:28 crc kubenswrapper[4675]: I0219 09:28:28.039974 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:28 crc kubenswrapper[4675]: I0219 09:28:28.088578 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:28 crc kubenswrapper[4675]: I0219 09:28:28.800659 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:28 crc kubenswrapper[4675]: I0219 09:28:28.847994 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bwfg6"] Feb 19 09:28:30 crc kubenswrapper[4675]: I0219 09:28:30.761413 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bwfg6" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="registry-server" containerID="cri-o://fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4" gracePeriod=2 Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.266692 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.439781 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-utilities\") pod \"45a88869-9ddb-4751-b44c-b035f974d06f\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.440329 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc9rs\" (UniqueName: \"kubernetes.io/projected/45a88869-9ddb-4751-b44c-b035f974d06f-kube-api-access-xc9rs\") pod \"45a88869-9ddb-4751-b44c-b035f974d06f\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.440486 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-catalog-content\") pod \"45a88869-9ddb-4751-b44c-b035f974d06f\" (UID: \"45a88869-9ddb-4751-b44c-b035f974d06f\") " Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.440861 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-utilities" (OuterVolumeSpecName: "utilities") pod "45a88869-9ddb-4751-b44c-b035f974d06f" (UID: "45a88869-9ddb-4751-b44c-b035f974d06f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.441192 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.448994 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45a88869-9ddb-4751-b44c-b035f974d06f-kube-api-access-xc9rs" (OuterVolumeSpecName: "kube-api-access-xc9rs") pod "45a88869-9ddb-4751-b44c-b035f974d06f" (UID: "45a88869-9ddb-4751-b44c-b035f974d06f"). InnerVolumeSpecName "kube-api-access-xc9rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.491677 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45a88869-9ddb-4751-b44c-b035f974d06f" (UID: "45a88869-9ddb-4751-b44c-b035f974d06f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.543071 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc9rs\" (UniqueName: \"kubernetes.io/projected/45a88869-9ddb-4751-b44c-b035f974d06f-kube-api-access-xc9rs\") on node \"crc\" DevicePath \"\"" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.543112 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a88869-9ddb-4751-b44c-b035f974d06f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.770711 4675 generic.go:334] "Generic (PLEG): container finished" podID="45a88869-9ddb-4751-b44c-b035f974d06f" containerID="fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4" exitCode=0 Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.770757 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwfg6" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.770765 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwfg6" event={"ID":"45a88869-9ddb-4751-b44c-b035f974d06f","Type":"ContainerDied","Data":"fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4"} Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.770799 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwfg6" event={"ID":"45a88869-9ddb-4751-b44c-b035f974d06f","Type":"ContainerDied","Data":"bb28e1b0d587eb0ac0841269c9c4ce04e598a0d079807dc3fb74f16413a48181"} Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.770817 4675 scope.go:117] "RemoveContainer" containerID="fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.804347 4675 scope.go:117] "RemoveContainer" containerID="e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.808728 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bwfg6"] Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.817527 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bwfg6"] Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.823833 4675 scope.go:117] "RemoveContainer" containerID="d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.865118 4675 scope.go:117] "RemoveContainer" containerID="fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4" Feb 19 09:28:31 crc kubenswrapper[4675]: E0219 09:28:31.865593 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4\": container with ID starting with fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4 not found: ID does not exist" containerID="fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.865627 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4"} err="failed to get container status \"fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4\": rpc error: code = NotFound desc = could not find container \"fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4\": container with ID starting with fd887022e8def8ddf91e97f20c1b240d69e84f50e45e607538161a2c60c45fe4 not found: ID does not exist" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.865664 4675 scope.go:117] "RemoveContainer" containerID="e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987" Feb 19 09:28:31 crc kubenswrapper[4675]: E0219 09:28:31.865893 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987\": container with ID starting with e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987 not found: ID does not exist" containerID="e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.865972 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987"} err="failed to get container status \"e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987\": rpc error: code = NotFound desc = could not find container \"e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987\": container with ID starting with e23db6f31ea530f7bf93ec4001d3b5de97d7d3afb0a36ffa80166f8b7ca64987 not found: ID does not exist" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.866043 4675 scope.go:117] "RemoveContainer" containerID="d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6" Feb 19 09:28:31 crc kubenswrapper[4675]: E0219 09:28:31.866274 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6\": container with ID starting with d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6 not found: ID does not exist" containerID="d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6" Feb 19 09:28:31 crc kubenswrapper[4675]: I0219 09:28:31.866296 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6"} err="failed to get container status \"d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6\": rpc error: code = NotFound desc = could not find container \"d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6\": container with ID starting with d332533ba50582a6142b80d6d9e62057ccb6194956a5ac1c537c74fc0ebcbdc6 not found: ID does not exist" Feb 19 09:28:33 crc kubenswrapper[4675]: I0219 09:28:33.112769 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" path="/var/lib/kubelet/pods/45a88869-9ddb-4751-b44c-b035f974d06f/volumes" Feb 19 09:28:42 crc kubenswrapper[4675]: I0219 09:28:42.103475 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:28:42 crc kubenswrapper[4675]: I0219 09:28:42.866503 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"e07709968bd5dbbc00b8de107ed3c4984dd88399a210a5b915106d896f7d1afe"} Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.152522 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk"] Feb 19 09:30:00 crc kubenswrapper[4675]: E0219 09:30:00.153769 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="extract-content" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.153789 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="extract-content" Feb 19 09:30:00 crc kubenswrapper[4675]: E0219 09:30:00.153807 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="extract-utilities" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.153814 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="extract-utilities" Feb 19 09:30:00 crc kubenswrapper[4675]: E0219 09:30:00.153837 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="registry-server" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.153844 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="registry-server" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.154095 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="45a88869-9ddb-4751-b44c-b035f974d06f" containerName="registry-server" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.154818 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.159374 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.159555 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.163108 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk"] Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.254203 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2263e98f-ab33-40b3-8e4d-4a3539072d82-config-volume\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.254297 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ltzz\" (UniqueName: \"kubernetes.io/projected/2263e98f-ab33-40b3-8e4d-4a3539072d82-kube-api-access-9ltzz\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.254439 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2263e98f-ab33-40b3-8e4d-4a3539072d82-secret-volume\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.356399 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2263e98f-ab33-40b3-8e4d-4a3539072d82-secret-volume\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.356503 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2263e98f-ab33-40b3-8e4d-4a3539072d82-config-volume\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.356563 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ltzz\" (UniqueName: \"kubernetes.io/projected/2263e98f-ab33-40b3-8e4d-4a3539072d82-kube-api-access-9ltzz\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.358003 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2263e98f-ab33-40b3-8e4d-4a3539072d82-config-volume\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.369690 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2263e98f-ab33-40b3-8e4d-4a3539072d82-secret-volume\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.373385 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ltzz\" (UniqueName: \"kubernetes.io/projected/2263e98f-ab33-40b3-8e4d-4a3539072d82-kube-api-access-9ltzz\") pod \"collect-profiles-29524890-tp2sk\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.515020 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:00 crc kubenswrapper[4675]: I0219 09:30:00.983486 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk"] Feb 19 09:30:01 crc kubenswrapper[4675]: I0219 09:30:01.539253 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" event={"ID":"2263e98f-ab33-40b3-8e4d-4a3539072d82","Type":"ContainerStarted","Data":"98fc06af0de31219ee1cf072f9dfd6d6d35e0e4d06a57b2f05cf0dd1b8889d43"} Feb 19 09:30:01 crc kubenswrapper[4675]: I0219 09:30:01.539318 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" event={"ID":"2263e98f-ab33-40b3-8e4d-4a3539072d82","Type":"ContainerStarted","Data":"f62326b3078ebbc778145d4bcee913b1a05a2f250202e05b4f05303cb606a78c"} Feb 19 09:30:01 crc kubenswrapper[4675]: I0219 09:30:01.560926 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" podStartSLOduration=1.560902089 podStartE2EDuration="1.560902089s" podCreationTimestamp="2026-02-19 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:30:01.555574657 +0000 UTC m=+2783.182664925" watchObservedRunningTime="2026-02-19 09:30:01.560902089 +0000 UTC m=+2783.187992357" Feb 19 09:30:02 crc kubenswrapper[4675]: I0219 09:30:02.547988 4675 generic.go:334] "Generic (PLEG): container finished" podID="2263e98f-ab33-40b3-8e4d-4a3539072d82" containerID="98fc06af0de31219ee1cf072f9dfd6d6d35e0e4d06a57b2f05cf0dd1b8889d43" exitCode=0 Feb 19 09:30:02 crc kubenswrapper[4675]: I0219 09:30:02.548035 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" event={"ID":"2263e98f-ab33-40b3-8e4d-4a3539072d82","Type":"ContainerDied","Data":"98fc06af0de31219ee1cf072f9dfd6d6d35e0e4d06a57b2f05cf0dd1b8889d43"} Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.065613 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.248055 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ltzz\" (UniqueName: \"kubernetes.io/projected/2263e98f-ab33-40b3-8e4d-4a3539072d82-kube-api-access-9ltzz\") pod \"2263e98f-ab33-40b3-8e4d-4a3539072d82\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.248149 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2263e98f-ab33-40b3-8e4d-4a3539072d82-config-volume\") pod \"2263e98f-ab33-40b3-8e4d-4a3539072d82\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.248222 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2263e98f-ab33-40b3-8e4d-4a3539072d82-secret-volume\") pod \"2263e98f-ab33-40b3-8e4d-4a3539072d82\" (UID: \"2263e98f-ab33-40b3-8e4d-4a3539072d82\") " Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.248811 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2263e98f-ab33-40b3-8e4d-4a3539072d82-config-volume" (OuterVolumeSpecName: "config-volume") pod "2263e98f-ab33-40b3-8e4d-4a3539072d82" (UID: "2263e98f-ab33-40b3-8e4d-4a3539072d82"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.250360 4675 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2263e98f-ab33-40b3-8e4d-4a3539072d82-config-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.257275 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2263e98f-ab33-40b3-8e4d-4a3539072d82-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2263e98f-ab33-40b3-8e4d-4a3539072d82" (UID: "2263e98f-ab33-40b3-8e4d-4a3539072d82"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.258350 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2263e98f-ab33-40b3-8e4d-4a3539072d82-kube-api-access-9ltzz" (OuterVolumeSpecName: "kube-api-access-9ltzz") pod "2263e98f-ab33-40b3-8e4d-4a3539072d82" (UID: "2263e98f-ab33-40b3-8e4d-4a3539072d82"). InnerVolumeSpecName "kube-api-access-9ltzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.352044 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ltzz\" (UniqueName: \"kubernetes.io/projected/2263e98f-ab33-40b3-8e4d-4a3539072d82-kube-api-access-9ltzz\") on node \"crc\" DevicePath \"\"" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.352077 4675 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2263e98f-ab33-40b3-8e4d-4a3539072d82-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.567809 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" event={"ID":"2263e98f-ab33-40b3-8e4d-4a3539072d82","Type":"ContainerDied","Data":"f62326b3078ebbc778145d4bcee913b1a05a2f250202e05b4f05303cb606a78c"} Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.567855 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f62326b3078ebbc778145d4bcee913b1a05a2f250202e05b4f05303cb606a78c" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.567924 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524890-tp2sk" Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.648341 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n"] Feb 19 09:30:04 crc kubenswrapper[4675]: I0219 09:30:04.667410 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524845-v9r5n"] Feb 19 09:30:05 crc kubenswrapper[4675]: I0219 09:30:05.118305 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd" path="/var/lib/kubelet/pods/e3b2dd0a-cf3b-49ad-bac1-18a0489ab6dd/volumes" Feb 19 09:30:58 crc kubenswrapper[4675]: I0219 09:30:58.750516 4675 scope.go:117] "RemoveContainer" containerID="34ac4eed0ec6be49c9adb1fff2a6b576a153d33e326eb80ac1eeb348009129c1" Feb 19 09:31:11 crc kubenswrapper[4675]: I0219 09:31:11.743946 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:31:11 crc kubenswrapper[4675]: I0219 09:31:11.744710 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.178150 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cs7rr"] Feb 19 09:31:31 crc kubenswrapper[4675]: E0219 09:31:31.179526 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2263e98f-ab33-40b3-8e4d-4a3539072d82" containerName="collect-profiles" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.179549 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2263e98f-ab33-40b3-8e4d-4a3539072d82" containerName="collect-profiles" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.179944 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="2263e98f-ab33-40b3-8e4d-4a3539072d82" containerName="collect-profiles" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.184175 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.190514 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cs7rr"] Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.281775 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhmxc\" (UniqueName: \"kubernetes.io/projected/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-kube-api-access-xhmxc\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.282280 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-catalog-content\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.282566 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-utilities\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.384183 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhmxc\" (UniqueName: \"kubernetes.io/projected/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-kube-api-access-xhmxc\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.384289 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-catalog-content\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.384391 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-utilities\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.385528 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-utilities\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.385563 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-catalog-content\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.415739 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhmxc\" (UniqueName: \"kubernetes.io/projected/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-kube-api-access-xhmxc\") pod \"redhat-operators-cs7rr\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.508858 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:31 crc kubenswrapper[4675]: I0219 09:31:31.993438 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cs7rr"] Feb 19 09:31:32 crc kubenswrapper[4675]: I0219 09:31:32.335754 4675 generic.go:334] "Generic (PLEG): container finished" podID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerID="65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d" exitCode=0 Feb 19 09:31:32 crc kubenswrapper[4675]: I0219 09:31:32.335806 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cs7rr" event={"ID":"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b","Type":"ContainerDied","Data":"65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d"} Feb 19 09:31:32 crc kubenswrapper[4675]: I0219 09:31:32.335839 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cs7rr" event={"ID":"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b","Type":"ContainerStarted","Data":"0569c54dfe36f5939447987ad938f0d045048b1c386cd2d90ec06fe03d3f2493"} Feb 19 09:31:34 crc kubenswrapper[4675]: I0219 09:31:34.362589 4675 generic.go:334] "Generic (PLEG): container finished" podID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerID="5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879" exitCode=0 Feb 19 09:31:34 crc kubenswrapper[4675]: I0219 09:31:34.362678 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cs7rr" event={"ID":"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b","Type":"ContainerDied","Data":"5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879"} Feb 19 09:31:35 crc kubenswrapper[4675]: I0219 09:31:35.374212 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cs7rr" event={"ID":"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b","Type":"ContainerStarted","Data":"b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b"} Feb 19 09:31:35 crc kubenswrapper[4675]: I0219 09:31:35.402568 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cs7rr" podStartSLOduration=1.941533715 podStartE2EDuration="4.402544152s" podCreationTimestamp="2026-02-19 09:31:31 +0000 UTC" firstStartedPulling="2026-02-19 09:31:32.338833837 +0000 UTC m=+2873.965924105" lastFinishedPulling="2026-02-19 09:31:34.799844274 +0000 UTC m=+2876.426934542" observedRunningTime="2026-02-19 09:31:35.39166277 +0000 UTC m=+2877.018753038" watchObservedRunningTime="2026-02-19 09:31:35.402544152 +0000 UTC m=+2877.029634410" Feb 19 09:31:41 crc kubenswrapper[4675]: I0219 09:31:41.509858 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:41 crc kubenswrapper[4675]: I0219 09:31:41.510522 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:41 crc kubenswrapper[4675]: I0219 09:31:41.561918 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:41 crc kubenswrapper[4675]: I0219 09:31:41.743976 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:31:41 crc kubenswrapper[4675]: I0219 09:31:41.744034 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:31:42 crc kubenswrapper[4675]: I0219 09:31:42.489915 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:42 crc kubenswrapper[4675]: I0219 09:31:42.538435 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cs7rr"] Feb 19 09:31:44 crc kubenswrapper[4675]: I0219 09:31:44.454447 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cs7rr" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="registry-server" containerID="cri-o://b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b" gracePeriod=2 Feb 19 09:31:44 crc kubenswrapper[4675]: I0219 09:31:44.954715 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.083178 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhmxc\" (UniqueName: \"kubernetes.io/projected/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-kube-api-access-xhmxc\") pod \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.083323 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-utilities\") pod \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.083583 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-catalog-content\") pod \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\" (UID: \"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b\") " Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.084329 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-utilities" (OuterVolumeSpecName: "utilities") pod "2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" (UID: "2a47bddf-e9f9-4d4b-ae02-760b4a14b88b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.090081 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-kube-api-access-xhmxc" (OuterVolumeSpecName: "kube-api-access-xhmxc") pod "2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" (UID: "2a47bddf-e9f9-4d4b-ae02-760b4a14b88b"). InnerVolumeSpecName "kube-api-access-xhmxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.185565 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhmxc\" (UniqueName: \"kubernetes.io/projected/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-kube-api-access-xhmxc\") on node \"crc\" DevicePath \"\"" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.185596 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.208723 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" (UID: "2a47bddf-e9f9-4d4b-ae02-760b4a14b88b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.287595 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.465020 4675 generic.go:334] "Generic (PLEG): container finished" podID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerID="b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b" exitCode=0 Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.465087 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cs7rr" event={"ID":"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b","Type":"ContainerDied","Data":"b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b"} Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.465123 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cs7rr" event={"ID":"2a47bddf-e9f9-4d4b-ae02-760b4a14b88b","Type":"ContainerDied","Data":"0569c54dfe36f5939447987ad938f0d045048b1c386cd2d90ec06fe03d3f2493"} Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.465141 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cs7rr" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.465160 4675 scope.go:117] "RemoveContainer" containerID="b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.485128 4675 scope.go:117] "RemoveContainer" containerID="5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.507550 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cs7rr"] Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.514825 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cs7rr"] Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.525801 4675 scope.go:117] "RemoveContainer" containerID="65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.560570 4675 scope.go:117] "RemoveContainer" containerID="b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b" Feb 19 09:31:45 crc kubenswrapper[4675]: E0219 09:31:45.561092 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b\": container with ID starting with b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b not found: ID does not exist" containerID="b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.561131 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b"} err="failed to get container status \"b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b\": rpc error: code = NotFound desc = could not find container \"b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b\": container with ID starting with b04592dd3557421f44d53f4488a0ab66ed0a2bb8937dded7d37fe5533d71ca4b not found: ID does not exist" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.561153 4675 scope.go:117] "RemoveContainer" containerID="5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879" Feb 19 09:31:45 crc kubenswrapper[4675]: E0219 09:31:45.561471 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879\": container with ID starting with 5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879 not found: ID does not exist" containerID="5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.561506 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879"} err="failed to get container status \"5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879\": rpc error: code = NotFound desc = could not find container \"5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879\": container with ID starting with 5ac27c5c06ae3f06a4e11dca6ee0b050bafa74d01017f68273621d035c9f1879 not found: ID does not exist" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.561528 4675 scope.go:117] "RemoveContainer" containerID="65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d" Feb 19 09:31:45 crc kubenswrapper[4675]: E0219 09:31:45.561782 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d\": container with ID starting with 65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d not found: ID does not exist" containerID="65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d" Feb 19 09:31:45 crc kubenswrapper[4675]: I0219 09:31:45.561808 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d"} err="failed to get container status \"65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d\": rpc error: code = NotFound desc = could not find container \"65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d\": container with ID starting with 65f34618a02099b5a466159a4da08f8ca859b17ebeeb9fe238f782d2838ffb2d not found: ID does not exist" Feb 19 09:31:47 crc kubenswrapper[4675]: I0219 09:31:47.113624 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" path="/var/lib/kubelet/pods/2a47bddf-e9f9-4d4b-ae02-760b4a14b88b/volumes" Feb 19 09:32:11 crc kubenswrapper[4675]: I0219 09:32:11.743799 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:32:11 crc kubenswrapper[4675]: I0219 09:32:11.744540 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:32:11 crc kubenswrapper[4675]: I0219 09:32:11.744611 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:32:11 crc kubenswrapper[4675]: I0219 09:32:11.745830 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e07709968bd5dbbc00b8de107ed3c4984dd88399a210a5b915106d896f7d1afe"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:32:11 crc kubenswrapper[4675]: I0219 09:32:11.745905 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://e07709968bd5dbbc00b8de107ed3c4984dd88399a210a5b915106d896f7d1afe" gracePeriod=600 Feb 19 09:32:12 crc kubenswrapper[4675]: I0219 09:32:12.696488 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="e07709968bd5dbbc00b8de107ed3c4984dd88399a210a5b915106d896f7d1afe" exitCode=0 Feb 19 09:32:12 crc kubenswrapper[4675]: I0219 09:32:12.696575 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"e07709968bd5dbbc00b8de107ed3c4984dd88399a210a5b915106d896f7d1afe"} Feb 19 09:32:12 crc kubenswrapper[4675]: I0219 09:32:12.697075 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687"} Feb 19 09:32:12 crc kubenswrapper[4675]: I0219 09:32:12.697120 4675 scope.go:117] "RemoveContainer" containerID="5a8118ab0c5e143e9dc8b14a114f80a0d5e8e77259f3c448ad5aa15dc4b0341b" Feb 19 09:34:41 crc kubenswrapper[4675]: I0219 09:34:41.744586 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:34:41 crc kubenswrapper[4675]: I0219 09:34:41.745570 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:35:11 crc kubenswrapper[4675]: I0219 09:35:11.743682 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:35:11 crc kubenswrapper[4675]: I0219 09:35:11.744295 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:35:41 crc kubenswrapper[4675]: I0219 09:35:41.743529 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:35:41 crc kubenswrapper[4675]: I0219 09:35:41.744311 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:35:41 crc kubenswrapper[4675]: I0219 09:35:41.744363 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:35:41 crc kubenswrapper[4675]: I0219 09:35:41.745177 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:35:41 crc kubenswrapper[4675]: I0219 09:35:41.745231 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" gracePeriod=600 Feb 19 09:35:41 crc kubenswrapper[4675]: E0219 09:35:41.871896 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:35:42 crc kubenswrapper[4675]: I0219 09:35:42.496668 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" exitCode=0 Feb 19 09:35:42 crc kubenswrapper[4675]: I0219 09:35:42.496701 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687"} Feb 19 09:35:42 crc kubenswrapper[4675]: I0219 09:35:42.496796 4675 scope.go:117] "RemoveContainer" containerID="e07709968bd5dbbc00b8de107ed3c4984dd88399a210a5b915106d896f7d1afe" Feb 19 09:35:42 crc kubenswrapper[4675]: I0219 09:35:42.497852 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:35:42 crc kubenswrapper[4675]: E0219 09:35:42.498312 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:35:56 crc kubenswrapper[4675]: I0219 09:35:56.103490 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:35:56 crc kubenswrapper[4675]: E0219 09:35:56.104486 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:36:11 crc kubenswrapper[4675]: I0219 09:36:11.103367 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:36:11 crc kubenswrapper[4675]: E0219 09:36:11.104688 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.577926 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nm9br"] Feb 19 09:36:12 crc kubenswrapper[4675]: E0219 09:36:12.578484 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="registry-server" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.578501 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="registry-server" Feb 19 09:36:12 crc kubenswrapper[4675]: E0219 09:36:12.578534 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="extract-utilities" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.578543 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="extract-utilities" Feb 19 09:36:12 crc kubenswrapper[4675]: E0219 09:36:12.578568 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="extract-content" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.578578 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="extract-content" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.578844 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a47bddf-e9f9-4d4b-ae02-760b4a14b88b" containerName="registry-server" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.580802 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.624933 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nm9br"] Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.689797 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-catalog-content\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.689945 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-utilities\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.689987 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmgj2\" (UniqueName: \"kubernetes.io/projected/2d2414db-725d-41cf-96d1-5dd27d35f234-kube-api-access-gmgj2\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.792174 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-catalog-content\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.792260 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-utilities\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.792282 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmgj2\" (UniqueName: \"kubernetes.io/projected/2d2414db-725d-41cf-96d1-5dd27d35f234-kube-api-access-gmgj2\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.792788 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-catalog-content\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.792857 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-utilities\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.818823 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmgj2\" (UniqueName: \"kubernetes.io/projected/2d2414db-725d-41cf-96d1-5dd27d35f234-kube-api-access-gmgj2\") pod \"certified-operators-nm9br\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:12 crc kubenswrapper[4675]: I0219 09:36:12.940730 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:13 crc kubenswrapper[4675]: I0219 09:36:13.447793 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nm9br"] Feb 19 09:36:13 crc kubenswrapper[4675]: I0219 09:36:13.781968 4675 generic.go:334] "Generic (PLEG): container finished" podID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerID="6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5" exitCode=0 Feb 19 09:36:13 crc kubenswrapper[4675]: I0219 09:36:13.782036 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm9br" event={"ID":"2d2414db-725d-41cf-96d1-5dd27d35f234","Type":"ContainerDied","Data":"6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5"} Feb 19 09:36:13 crc kubenswrapper[4675]: I0219 09:36:13.782836 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm9br" event={"ID":"2d2414db-725d-41cf-96d1-5dd27d35f234","Type":"ContainerStarted","Data":"ea7a8c3fc1fdee10c269036c20b9da07a61508b280f56aa5ce7125dc377380b4"} Feb 19 09:36:13 crc kubenswrapper[4675]: I0219 09:36:13.784135 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:36:14 crc kubenswrapper[4675]: I0219 09:36:14.793534 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm9br" event={"ID":"2d2414db-725d-41cf-96d1-5dd27d35f234","Type":"ContainerStarted","Data":"78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d"} Feb 19 09:36:15 crc kubenswrapper[4675]: I0219 09:36:15.803956 4675 generic.go:334] "Generic (PLEG): container finished" podID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerID="78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d" exitCode=0 Feb 19 09:36:15 crc kubenswrapper[4675]: I0219 09:36:15.804031 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm9br" event={"ID":"2d2414db-725d-41cf-96d1-5dd27d35f234","Type":"ContainerDied","Data":"78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d"} Feb 19 09:36:16 crc kubenswrapper[4675]: I0219 09:36:16.818062 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm9br" event={"ID":"2d2414db-725d-41cf-96d1-5dd27d35f234","Type":"ContainerStarted","Data":"2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7"} Feb 19 09:36:16 crc kubenswrapper[4675]: I0219 09:36:16.838281 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nm9br" podStartSLOduration=2.41791247 podStartE2EDuration="4.838255986s" podCreationTimestamp="2026-02-19 09:36:12 +0000 UTC" firstStartedPulling="2026-02-19 09:36:13.783839301 +0000 UTC m=+3155.410929579" lastFinishedPulling="2026-02-19 09:36:16.204182827 +0000 UTC m=+3157.831273095" observedRunningTime="2026-02-19 09:36:16.835118332 +0000 UTC m=+3158.462208620" watchObservedRunningTime="2026-02-19 09:36:16.838255986 +0000 UTC m=+3158.465346254" Feb 19 09:36:19 crc kubenswrapper[4675]: I0219 09:36:19.946419 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5hp4j"] Feb 19 09:36:19 crc kubenswrapper[4675]: I0219 09:36:19.949352 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:19 crc kubenswrapper[4675]: I0219 09:36:19.976704 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hp4j"] Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.077068 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-utilities\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.077984 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxbb5\" (UniqueName: \"kubernetes.io/projected/fea64196-4e16-43d3-b35e-52001a80f318-kube-api-access-hxbb5\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.078226 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-catalog-content\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.181066 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-utilities\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.181260 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxbb5\" (UniqueName: \"kubernetes.io/projected/fea64196-4e16-43d3-b35e-52001a80f318-kube-api-access-hxbb5\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.181305 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-catalog-content\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.182024 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-catalog-content\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.182973 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-utilities\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.208407 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxbb5\" (UniqueName: \"kubernetes.io/projected/fea64196-4e16-43d3-b35e-52001a80f318-kube-api-access-hxbb5\") pod \"redhat-marketplace-5hp4j\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.282274 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:20 crc kubenswrapper[4675]: I0219 09:36:20.849473 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hp4j"] Feb 19 09:36:20 crc kubenswrapper[4675]: W0219 09:36:20.855311 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfea64196_4e16_43d3_b35e_52001a80f318.slice/crio-fb5b9035825d1513087d86340b0a869f5d4f8c5eeea3f0d65310606571ba2f09 WatchSource:0}: Error finding container fb5b9035825d1513087d86340b0a869f5d4f8c5eeea3f0d65310606571ba2f09: Status 404 returned error can't find the container with id fb5b9035825d1513087d86340b0a869f5d4f8c5eeea3f0d65310606571ba2f09 Feb 19 09:36:21 crc kubenswrapper[4675]: I0219 09:36:21.862767 4675 generic.go:334] "Generic (PLEG): container finished" podID="fea64196-4e16-43d3-b35e-52001a80f318" containerID="82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b" exitCode=0 Feb 19 09:36:21 crc kubenswrapper[4675]: I0219 09:36:21.863198 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hp4j" event={"ID":"fea64196-4e16-43d3-b35e-52001a80f318","Type":"ContainerDied","Data":"82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b"} Feb 19 09:36:21 crc kubenswrapper[4675]: I0219 09:36:21.863228 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hp4j" event={"ID":"fea64196-4e16-43d3-b35e-52001a80f318","Type":"ContainerStarted","Data":"fb5b9035825d1513087d86340b0a869f5d4f8c5eeea3f0d65310606571ba2f09"} Feb 19 09:36:22 crc kubenswrapper[4675]: I0219 09:36:22.103991 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:36:22 crc kubenswrapper[4675]: E0219 09:36:22.104527 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:36:22 crc kubenswrapper[4675]: I0219 09:36:22.873871 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hp4j" event={"ID":"fea64196-4e16-43d3-b35e-52001a80f318","Type":"ContainerStarted","Data":"cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb"} Feb 19 09:36:22 crc kubenswrapper[4675]: I0219 09:36:22.941161 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:22 crc kubenswrapper[4675]: I0219 09:36:22.941481 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:22 crc kubenswrapper[4675]: I0219 09:36:22.987927 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:23 crc kubenswrapper[4675]: I0219 09:36:23.883020 4675 generic.go:334] "Generic (PLEG): container finished" podID="fea64196-4e16-43d3-b35e-52001a80f318" containerID="cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb" exitCode=0 Feb 19 09:36:23 crc kubenswrapper[4675]: I0219 09:36:23.883140 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hp4j" event={"ID":"fea64196-4e16-43d3-b35e-52001a80f318","Type":"ContainerDied","Data":"cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb"} Feb 19 09:36:23 crc kubenswrapper[4675]: I0219 09:36:23.949329 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:24 crc kubenswrapper[4675]: I0219 09:36:24.894921 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hp4j" event={"ID":"fea64196-4e16-43d3-b35e-52001a80f318","Type":"ContainerStarted","Data":"0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9"} Feb 19 09:36:24 crc kubenswrapper[4675]: I0219 09:36:24.927138 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5hp4j" podStartSLOduration=3.562656462 podStartE2EDuration="5.92710851s" podCreationTimestamp="2026-02-19 09:36:19 +0000 UTC" firstStartedPulling="2026-02-19 09:36:21.867538676 +0000 UTC m=+3163.494628944" lastFinishedPulling="2026-02-19 09:36:24.231990734 +0000 UTC m=+3165.859080992" observedRunningTime="2026-02-19 09:36:24.917011519 +0000 UTC m=+3166.544101797" watchObservedRunningTime="2026-02-19 09:36:24.92710851 +0000 UTC m=+3166.554198788" Feb 19 09:36:25 crc kubenswrapper[4675]: I0219 09:36:25.334610 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nm9br"] Feb 19 09:36:26 crc kubenswrapper[4675]: I0219 09:36:26.912115 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nm9br" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="registry-server" containerID="cri-o://2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7" gracePeriod=2 Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.392868 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.443598 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-utilities\") pod \"2d2414db-725d-41cf-96d1-5dd27d35f234\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.444041 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-catalog-content\") pod \"2d2414db-725d-41cf-96d1-5dd27d35f234\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.444107 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmgj2\" (UniqueName: \"kubernetes.io/projected/2d2414db-725d-41cf-96d1-5dd27d35f234-kube-api-access-gmgj2\") pod \"2d2414db-725d-41cf-96d1-5dd27d35f234\" (UID: \"2d2414db-725d-41cf-96d1-5dd27d35f234\") " Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.444999 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-utilities" (OuterVolumeSpecName: "utilities") pod "2d2414db-725d-41cf-96d1-5dd27d35f234" (UID: "2d2414db-725d-41cf-96d1-5dd27d35f234"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.453804 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d2414db-725d-41cf-96d1-5dd27d35f234-kube-api-access-gmgj2" (OuterVolumeSpecName: "kube-api-access-gmgj2") pod "2d2414db-725d-41cf-96d1-5dd27d35f234" (UID: "2d2414db-725d-41cf-96d1-5dd27d35f234"). InnerVolumeSpecName "kube-api-access-gmgj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.498383 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d2414db-725d-41cf-96d1-5dd27d35f234" (UID: "2d2414db-725d-41cf-96d1-5dd27d35f234"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.547553 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.547632 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2414db-725d-41cf-96d1-5dd27d35f234-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.547703 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmgj2\" (UniqueName: \"kubernetes.io/projected/2d2414db-725d-41cf-96d1-5dd27d35f234-kube-api-access-gmgj2\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.927118 4675 generic.go:334] "Generic (PLEG): container finished" podID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerID="2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7" exitCode=0 Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.927157 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nm9br" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.927186 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm9br" event={"ID":"2d2414db-725d-41cf-96d1-5dd27d35f234","Type":"ContainerDied","Data":"2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7"} Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.927890 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nm9br" event={"ID":"2d2414db-725d-41cf-96d1-5dd27d35f234","Type":"ContainerDied","Data":"ea7a8c3fc1fdee10c269036c20b9da07a61508b280f56aa5ce7125dc377380b4"} Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.927926 4675 scope.go:117] "RemoveContainer" containerID="2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.965157 4675 scope.go:117] "RemoveContainer" containerID="78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d" Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.973055 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nm9br"] Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.985115 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nm9br"] Feb 19 09:36:27 crc kubenswrapper[4675]: I0219 09:36:27.995146 4675 scope.go:117] "RemoveContainer" containerID="6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5" Feb 19 09:36:28 crc kubenswrapper[4675]: I0219 09:36:28.052613 4675 scope.go:117] "RemoveContainer" containerID="2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7" Feb 19 09:36:28 crc kubenswrapper[4675]: E0219 09:36:28.053323 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7\": container with ID starting with 2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7 not found: ID does not exist" containerID="2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7" Feb 19 09:36:28 crc kubenswrapper[4675]: I0219 09:36:28.053354 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7"} err="failed to get container status \"2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7\": rpc error: code = NotFound desc = could not find container \"2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7\": container with ID starting with 2e1f55a8bfe01bffd1e7cebfdac73b4d054b82d2d1899105c8edab837c129ab7 not found: ID does not exist" Feb 19 09:36:28 crc kubenswrapper[4675]: I0219 09:36:28.053376 4675 scope.go:117] "RemoveContainer" containerID="78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d" Feb 19 09:36:28 crc kubenswrapper[4675]: E0219 09:36:28.053919 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d\": container with ID starting with 78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d not found: ID does not exist" containerID="78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d" Feb 19 09:36:28 crc kubenswrapper[4675]: I0219 09:36:28.053942 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d"} err="failed to get container status \"78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d\": rpc error: code = NotFound desc = could not find container \"78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d\": container with ID starting with 78057b285b9b19a46ce0719776e215c287ffe013caf69951fd7895b8175aa30d not found: ID does not exist" Feb 19 09:36:28 crc kubenswrapper[4675]: I0219 09:36:28.053954 4675 scope.go:117] "RemoveContainer" containerID="6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5" Feb 19 09:36:28 crc kubenswrapper[4675]: E0219 09:36:28.054421 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5\": container with ID starting with 6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5 not found: ID does not exist" containerID="6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5" Feb 19 09:36:28 crc kubenswrapper[4675]: I0219 09:36:28.054471 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5"} err="failed to get container status \"6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5\": rpc error: code = NotFound desc = could not find container \"6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5\": container with ID starting with 6b56b8feca2179eb3796d7fd5c6fff36936fa47ccf6f6a0eecca97bf2b7cabd5 not found: ID does not exist" Feb 19 09:36:29 crc kubenswrapper[4675]: I0219 09:36:29.117268 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" path="/var/lib/kubelet/pods/2d2414db-725d-41cf-96d1-5dd27d35f234/volumes" Feb 19 09:36:30 crc kubenswrapper[4675]: I0219 09:36:30.284115 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:30 crc kubenswrapper[4675]: I0219 09:36:30.284161 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:30 crc kubenswrapper[4675]: I0219 09:36:30.338316 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:31 crc kubenswrapper[4675]: I0219 09:36:31.005629 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:31 crc kubenswrapper[4675]: I0219 09:36:31.534468 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hp4j"] Feb 19 09:36:32 crc kubenswrapper[4675]: I0219 09:36:32.972418 4675 generic.go:334] "Generic (PLEG): container finished" podID="5006db86-ae9d-4dbc-871f-4bc18f3a43f6" containerID="b10817aa3eeed3cee7c7e92e420ba77914b0ec9c562c157f9cb0393cd4b7034b" exitCode=0 Feb 19 09:36:32 crc kubenswrapper[4675]: I0219 09:36:32.972515 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5006db86-ae9d-4dbc-871f-4bc18f3a43f6","Type":"ContainerDied","Data":"b10817aa3eeed3cee7c7e92e420ba77914b0ec9c562c157f9cb0393cd4b7034b"} Feb 19 09:36:32 crc kubenswrapper[4675]: I0219 09:36:32.973010 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5hp4j" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="registry-server" containerID="cri-o://0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9" gracePeriod=2 Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.464506 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.570099 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxbb5\" (UniqueName: \"kubernetes.io/projected/fea64196-4e16-43d3-b35e-52001a80f318-kube-api-access-hxbb5\") pod \"fea64196-4e16-43d3-b35e-52001a80f318\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.570174 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-catalog-content\") pod \"fea64196-4e16-43d3-b35e-52001a80f318\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.570244 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-utilities\") pod \"fea64196-4e16-43d3-b35e-52001a80f318\" (UID: \"fea64196-4e16-43d3-b35e-52001a80f318\") " Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.571052 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-utilities" (OuterVolumeSpecName: "utilities") pod "fea64196-4e16-43d3-b35e-52001a80f318" (UID: "fea64196-4e16-43d3-b35e-52001a80f318"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.582011 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea64196-4e16-43d3-b35e-52001a80f318-kube-api-access-hxbb5" (OuterVolumeSpecName: "kube-api-access-hxbb5") pod "fea64196-4e16-43d3-b35e-52001a80f318" (UID: "fea64196-4e16-43d3-b35e-52001a80f318"). InnerVolumeSpecName "kube-api-access-hxbb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.599299 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fea64196-4e16-43d3-b35e-52001a80f318" (UID: "fea64196-4e16-43d3-b35e-52001a80f318"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.673064 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxbb5\" (UniqueName: \"kubernetes.io/projected/fea64196-4e16-43d3-b35e-52001a80f318-kube-api-access-hxbb5\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.673097 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.673106 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea64196-4e16-43d3-b35e-52001a80f318-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.986568 4675 generic.go:334] "Generic (PLEG): container finished" podID="fea64196-4e16-43d3-b35e-52001a80f318" containerID="0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9" exitCode=0 Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.986726 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hp4j" event={"ID":"fea64196-4e16-43d3-b35e-52001a80f318","Type":"ContainerDied","Data":"0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9"} Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.986757 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hp4j" Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.986789 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hp4j" event={"ID":"fea64196-4e16-43d3-b35e-52001a80f318","Type":"ContainerDied","Data":"fb5b9035825d1513087d86340b0a869f5d4f8c5eeea3f0d65310606571ba2f09"} Feb 19 09:36:33 crc kubenswrapper[4675]: I0219 09:36:33.986815 4675 scope.go:117] "RemoveContainer" containerID="0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.059863 4675 scope.go:117] "RemoveContainer" containerID="cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.074958 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hp4j"] Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.097681 4675 scope.go:117] "RemoveContainer" containerID="82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.103126 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:36:34 crc kubenswrapper[4675]: E0219 09:36:34.103567 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.109081 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hp4j"] Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.158862 4675 scope.go:117] "RemoveContainer" containerID="0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9" Feb 19 09:36:34 crc kubenswrapper[4675]: E0219 09:36:34.160873 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9\": container with ID starting with 0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9 not found: ID does not exist" containerID="0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.160927 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9"} err="failed to get container status \"0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9\": rpc error: code = NotFound desc = could not find container \"0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9\": container with ID starting with 0f45ee04eaf64c5639d5a622fce1f9473a5e9044454d6ee7d3edf9e4952094f9 not found: ID does not exist" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.160956 4675 scope.go:117] "RemoveContainer" containerID="cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb" Feb 19 09:36:34 crc kubenswrapper[4675]: E0219 09:36:34.164337 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb\": container with ID starting with cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb not found: ID does not exist" containerID="cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.164369 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb"} err="failed to get container status \"cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb\": rpc error: code = NotFound desc = could not find container \"cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb\": container with ID starting with cb8a24849a95a572fea8eacd554d295954fbd59ed51c27962ae2fda8c39f59eb not found: ID does not exist" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.164391 4675 scope.go:117] "RemoveContainer" containerID="82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b" Feb 19 09:36:34 crc kubenswrapper[4675]: E0219 09:36:34.165416 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b\": container with ID starting with 82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b not found: ID does not exist" containerID="82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.165434 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b"} err="failed to get container status \"82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b\": rpc error: code = NotFound desc = could not find container \"82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b\": container with ID starting with 82c2bd702339a3d13d17e5ab426ded7bb962c37ef3c05da0af722c14d4574d7b not found: ID does not exist" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.376096 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493270 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ca-certs\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493477 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-config-data\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493501 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scl6w\" (UniqueName: \"kubernetes.io/projected/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-kube-api-access-scl6w\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493532 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493566 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493618 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-temporary\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493655 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-workdir\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493687 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config-secret\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.493751 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ssh-key\") pod \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\" (UID: \"5006db86-ae9d-4dbc-871f-4bc18f3a43f6\") " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.495150 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.495164 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-config-data" (OuterVolumeSpecName: "config-data") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.499652 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.503149 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-kube-api-access-scl6w" (OuterVolumeSpecName: "kube-api-access-scl6w") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "kube-api-access-scl6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.506396 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.525870 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.528793 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.535620 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.547168 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5006db86-ae9d-4dbc-871f-4bc18f3a43f6" (UID: "5006db86-ae9d-4dbc-871f-4bc18f3a43f6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.596941 4675 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.596971 4675 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-config-data\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.596982 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scl6w\" (UniqueName: \"kubernetes.io/projected/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-kube-api-access-scl6w\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.596992 4675 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.597032 4675 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.597043 4675 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.597054 4675 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.597066 4675 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.609104 4675 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5006db86-ae9d-4dbc-871f-4bc18f3a43f6-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.630591 4675 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.711354 4675 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.996970 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5006db86-ae9d-4dbc-871f-4bc18f3a43f6","Type":"ContainerDied","Data":"4e1d7d6533475cf6e7c0cf274177ea15c311b225b3eb205fcdc0c64a072008f2"} Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.997009 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e1d7d6533475cf6e7c0cf274177ea15c311b225b3eb205fcdc0c64a072008f2" Feb 19 09:36:34 crc kubenswrapper[4675]: I0219 09:36:34.997051 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 19 09:36:35 crc kubenswrapper[4675]: I0219 09:36:35.121910 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fea64196-4e16-43d3-b35e-52001a80f318" path="/var/lib/kubelet/pods/fea64196-4e16-43d3-b35e-52001a80f318/volumes" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.380842 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 19 09:36:42 crc kubenswrapper[4675]: E0219 09:36:42.382343 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5006db86-ae9d-4dbc-871f-4bc18f3a43f6" containerName="tempest-tests-tempest-tests-runner" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382365 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="5006db86-ae9d-4dbc-871f-4bc18f3a43f6" containerName="tempest-tests-tempest-tests-runner" Feb 19 09:36:42 crc kubenswrapper[4675]: E0219 09:36:42.382378 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="registry-server" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382386 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="registry-server" Feb 19 09:36:42 crc kubenswrapper[4675]: E0219 09:36:42.382407 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="registry-server" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382417 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="registry-server" Feb 19 09:36:42 crc kubenswrapper[4675]: E0219 09:36:42.382446 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="extract-content" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382455 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="extract-content" Feb 19 09:36:42 crc kubenswrapper[4675]: E0219 09:36:42.382479 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="extract-utilities" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382488 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="extract-utilities" Feb 19 09:36:42 crc kubenswrapper[4675]: E0219 09:36:42.382505 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="extract-content" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382514 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="extract-content" Feb 19 09:36:42 crc kubenswrapper[4675]: E0219 09:36:42.382531 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="extract-utilities" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382540 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="extract-utilities" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382822 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d2414db-725d-41cf-96d1-5dd27d35f234" containerName="registry-server" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382845 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="fea64196-4e16-43d3-b35e-52001a80f318" containerName="registry-server" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.382879 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="5006db86-ae9d-4dbc-871f-4bc18f3a43f6" containerName="tempest-tests-tempest-tests-runner" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.383692 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.385679 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-h67kp" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.389302 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.479162 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b255b445-e2d1-4d12-8a53-34e76f18dede\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.479284 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttjz2\" (UniqueName: \"kubernetes.io/projected/b255b445-e2d1-4d12-8a53-34e76f18dede-kube-api-access-ttjz2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b255b445-e2d1-4d12-8a53-34e76f18dede\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.581177 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttjz2\" (UniqueName: \"kubernetes.io/projected/b255b445-e2d1-4d12-8a53-34e76f18dede-kube-api-access-ttjz2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b255b445-e2d1-4d12-8a53-34e76f18dede\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.581346 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b255b445-e2d1-4d12-8a53-34e76f18dede\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.581878 4675 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b255b445-e2d1-4d12-8a53-34e76f18dede\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.612461 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttjz2\" (UniqueName: \"kubernetes.io/projected/b255b445-e2d1-4d12-8a53-34e76f18dede-kube-api-access-ttjz2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b255b445-e2d1-4d12-8a53-34e76f18dede\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.614249 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b255b445-e2d1-4d12-8a53-34e76f18dede\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:42 crc kubenswrapper[4675]: I0219 09:36:42.718939 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 19 09:36:43 crc kubenswrapper[4675]: I0219 09:36:43.176618 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 19 09:36:44 crc kubenswrapper[4675]: I0219 09:36:44.086438 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b255b445-e2d1-4d12-8a53-34e76f18dede","Type":"ContainerStarted","Data":"58f7c80f2a33bbbe2fb8e6be0ab15a5ff1cfedd3dc54b78ec4c553ecb42ae7f9"} Feb 19 09:36:45 crc kubenswrapper[4675]: I0219 09:36:45.095577 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b255b445-e2d1-4d12-8a53-34e76f18dede","Type":"ContainerStarted","Data":"2207eb78ebb13ff41e3c298c447bc469b9e672d49badff35e87e35d006540053"} Feb 19 09:36:45 crc kubenswrapper[4675]: I0219 09:36:45.112321 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.396641956 podStartE2EDuration="3.112299292s" podCreationTimestamp="2026-02-19 09:36:42 +0000 UTC" firstStartedPulling="2026-02-19 09:36:43.18040686 +0000 UTC m=+3184.807497128" lastFinishedPulling="2026-02-19 09:36:43.896064206 +0000 UTC m=+3185.523154464" observedRunningTime="2026-02-19 09:36:45.108258653 +0000 UTC m=+3186.735348931" watchObservedRunningTime="2026-02-19 09:36:45.112299292 +0000 UTC m=+3186.739389570" Feb 19 09:36:49 crc kubenswrapper[4675]: I0219 09:36:49.110005 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:36:49 crc kubenswrapper[4675]: E0219 09:36:49.111069 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:37:02 crc kubenswrapper[4675]: I0219 09:37:02.104242 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:37:02 crc kubenswrapper[4675]: E0219 09:37:02.107479 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:37:06 crc kubenswrapper[4675]: I0219 09:37:06.763245 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-l245p/must-gather-qz4d2"] Feb 19 09:37:06 crc kubenswrapper[4675]: I0219 09:37:06.765560 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:06 crc kubenswrapper[4675]: I0219 09:37:06.773030 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-l245p"/"openshift-service-ca.crt" Feb 19 09:37:06 crc kubenswrapper[4675]: I0219 09:37:06.773269 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-l245p"/"kube-root-ca.crt" Feb 19 09:37:06 crc kubenswrapper[4675]: I0219 09:37:06.773391 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-l245p/must-gather-qz4d2"] Feb 19 09:37:06 crc kubenswrapper[4675]: I0219 09:37:06.926882 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0b7669be-8127-4b9b-a79d-b94ac608cb55-must-gather-output\") pod \"must-gather-qz4d2\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:06 crc kubenswrapper[4675]: I0219 09:37:06.926933 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6hqj\" (UniqueName: \"kubernetes.io/projected/0b7669be-8127-4b9b-a79d-b94ac608cb55-kube-api-access-x6hqj\") pod \"must-gather-qz4d2\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:07 crc kubenswrapper[4675]: I0219 09:37:07.029257 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0b7669be-8127-4b9b-a79d-b94ac608cb55-must-gather-output\") pod \"must-gather-qz4d2\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:07 crc kubenswrapper[4675]: I0219 09:37:07.029578 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6hqj\" (UniqueName: \"kubernetes.io/projected/0b7669be-8127-4b9b-a79d-b94ac608cb55-kube-api-access-x6hqj\") pod \"must-gather-qz4d2\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:07 crc kubenswrapper[4675]: I0219 09:37:07.030033 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0b7669be-8127-4b9b-a79d-b94ac608cb55-must-gather-output\") pod \"must-gather-qz4d2\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:07 crc kubenswrapper[4675]: I0219 09:37:07.048769 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6hqj\" (UniqueName: \"kubernetes.io/projected/0b7669be-8127-4b9b-a79d-b94ac608cb55-kube-api-access-x6hqj\") pod \"must-gather-qz4d2\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:07 crc kubenswrapper[4675]: I0219 09:37:07.093106 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:37:07 crc kubenswrapper[4675]: I0219 09:37:07.658041 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-l245p/must-gather-qz4d2"] Feb 19 09:37:07 crc kubenswrapper[4675]: W0219 09:37:07.661403 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b7669be_8127_4b9b_a79d_b94ac608cb55.slice/crio-ec9f307558fbc0279c43eef4f024afb946174cfd2a44afaf4883147f94c7a7ca WatchSource:0}: Error finding container ec9f307558fbc0279c43eef4f024afb946174cfd2a44afaf4883147f94c7a7ca: Status 404 returned error can't find the container with id ec9f307558fbc0279c43eef4f024afb946174cfd2a44afaf4883147f94c7a7ca Feb 19 09:37:08 crc kubenswrapper[4675]: I0219 09:37:08.344701 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/must-gather-qz4d2" event={"ID":"0b7669be-8127-4b9b-a79d-b94ac608cb55","Type":"ContainerStarted","Data":"ec9f307558fbc0279c43eef4f024afb946174cfd2a44afaf4883147f94c7a7ca"} Feb 19 09:37:14 crc kubenswrapper[4675]: I0219 09:37:14.397833 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/must-gather-qz4d2" event={"ID":"0b7669be-8127-4b9b-a79d-b94ac608cb55","Type":"ContainerStarted","Data":"0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae"} Feb 19 09:37:15 crc kubenswrapper[4675]: I0219 09:37:15.411842 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/must-gather-qz4d2" event={"ID":"0b7669be-8127-4b9b-a79d-b94ac608cb55","Type":"ContainerStarted","Data":"cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86"} Feb 19 09:37:15 crc kubenswrapper[4675]: I0219 09:37:15.429905 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-l245p/must-gather-qz4d2" podStartSLOduration=3.114697041 podStartE2EDuration="9.429886634s" podCreationTimestamp="2026-02-19 09:37:06 +0000 UTC" firstStartedPulling="2026-02-19 09:37:07.663580788 +0000 UTC m=+3209.290671056" lastFinishedPulling="2026-02-19 09:37:13.978770381 +0000 UTC m=+3215.605860649" observedRunningTime="2026-02-19 09:37:15.42600601 +0000 UTC m=+3217.053096278" watchObservedRunningTime="2026-02-19 09:37:15.429886634 +0000 UTC m=+3217.056976892" Feb 19 09:37:16 crc kubenswrapper[4675]: I0219 09:37:16.103072 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:37:16 crc kubenswrapper[4675]: E0219 09:37:16.103704 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:37:17 crc kubenswrapper[4675]: E0219 09:37:17.465035 4675 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.158:38988->38.102.83.158:42175: write tcp 38.102.83.158:38988->38.102.83.158:42175: write: broken pipe Feb 19 09:37:17 crc kubenswrapper[4675]: I0219 09:37:17.993095 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-l245p/crc-debug-vcd2j"] Feb 19 09:37:17 crc kubenswrapper[4675]: I0219 09:37:17.994782 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:17 crc kubenswrapper[4675]: I0219 09:37:17.997369 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-l245p"/"default-dockercfg-n66vl" Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.174417 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfmd4\" (UniqueName: \"kubernetes.io/projected/f77b2245-7b94-4691-b1cd-f611008439bb-kube-api-access-dfmd4\") pod \"crc-debug-vcd2j\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.174574 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f77b2245-7b94-4691-b1cd-f611008439bb-host\") pod \"crc-debug-vcd2j\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.277246 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfmd4\" (UniqueName: \"kubernetes.io/projected/f77b2245-7b94-4691-b1cd-f611008439bb-kube-api-access-dfmd4\") pod \"crc-debug-vcd2j\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.277466 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f77b2245-7b94-4691-b1cd-f611008439bb-host\") pod \"crc-debug-vcd2j\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.280430 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f77b2245-7b94-4691-b1cd-f611008439bb-host\") pod \"crc-debug-vcd2j\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.304276 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfmd4\" (UniqueName: \"kubernetes.io/projected/f77b2245-7b94-4691-b1cd-f611008439bb-kube-api-access-dfmd4\") pod \"crc-debug-vcd2j\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.323132 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:37:18 crc kubenswrapper[4675]: W0219 09:37:18.351746 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf77b2245_7b94_4691_b1cd_f611008439bb.slice/crio-e9fd17ab21e62652f9bb3ef164d5b93868322b7281b3ebc807d8d70aece2c2a0 WatchSource:0}: Error finding container e9fd17ab21e62652f9bb3ef164d5b93868322b7281b3ebc807d8d70aece2c2a0: Status 404 returned error can't find the container with id e9fd17ab21e62652f9bb3ef164d5b93868322b7281b3ebc807d8d70aece2c2a0 Feb 19 09:37:18 crc kubenswrapper[4675]: I0219 09:37:18.440015 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-vcd2j" event={"ID":"f77b2245-7b94-4691-b1cd-f611008439bb","Type":"ContainerStarted","Data":"e9fd17ab21e62652f9bb3ef164d5b93868322b7281b3ebc807d8d70aece2c2a0"} Feb 19 09:37:28 crc kubenswrapper[4675]: I0219 09:37:28.103023 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:37:28 crc kubenswrapper[4675]: E0219 09:37:28.103971 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:37:31 crc kubenswrapper[4675]: I0219 09:37:31.552402 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-vcd2j" event={"ID":"f77b2245-7b94-4691-b1cd-f611008439bb","Type":"ContainerStarted","Data":"86820b4f1b3cebbf94aba162f6db04d3638997282615c85e44e8ae3973e85b31"} Feb 19 09:37:31 crc kubenswrapper[4675]: I0219 09:37:31.567067 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-l245p/crc-debug-vcd2j" podStartSLOduration=2.008889912 podStartE2EDuration="14.567049084s" podCreationTimestamp="2026-02-19 09:37:17 +0000 UTC" firstStartedPulling="2026-02-19 09:37:18.353930434 +0000 UTC m=+3219.981020702" lastFinishedPulling="2026-02-19 09:37:30.912089606 +0000 UTC m=+3232.539179874" observedRunningTime="2026-02-19 09:37:31.563434747 +0000 UTC m=+3233.190525015" watchObservedRunningTime="2026-02-19 09:37:31.567049084 +0000 UTC m=+3233.194139352" Feb 19 09:37:43 crc kubenswrapper[4675]: I0219 09:37:43.103574 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:37:43 crc kubenswrapper[4675]: E0219 09:37:43.104785 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:37:55 crc kubenswrapper[4675]: I0219 09:37:55.103502 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:37:55 crc kubenswrapper[4675]: E0219 09:37:55.104524 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:38:07 crc kubenswrapper[4675]: I0219 09:38:07.103793 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:38:07 crc kubenswrapper[4675]: E0219 09:38:07.104791 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:38:08 crc kubenswrapper[4675]: I0219 09:38:08.209159 4675 generic.go:334] "Generic (PLEG): container finished" podID="f77b2245-7b94-4691-b1cd-f611008439bb" containerID="86820b4f1b3cebbf94aba162f6db04d3638997282615c85e44e8ae3973e85b31" exitCode=0 Feb 19 09:38:08 crc kubenswrapper[4675]: I0219 09:38:08.209240 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-vcd2j" event={"ID":"f77b2245-7b94-4691-b1cd-f611008439bb","Type":"ContainerDied","Data":"86820b4f1b3cebbf94aba162f6db04d3638997282615c85e44e8ae3973e85b31"} Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.368389 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.403435 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-l245p/crc-debug-vcd2j"] Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.413123 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-l245p/crc-debug-vcd2j"] Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.488292 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfmd4\" (UniqueName: \"kubernetes.io/projected/f77b2245-7b94-4691-b1cd-f611008439bb-kube-api-access-dfmd4\") pod \"f77b2245-7b94-4691-b1cd-f611008439bb\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.489463 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f77b2245-7b94-4691-b1cd-f611008439bb-host\") pod \"f77b2245-7b94-4691-b1cd-f611008439bb\" (UID: \"f77b2245-7b94-4691-b1cd-f611008439bb\") " Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.489582 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f77b2245-7b94-4691-b1cd-f611008439bb-host" (OuterVolumeSpecName: "host") pod "f77b2245-7b94-4691-b1cd-f611008439bb" (UID: "f77b2245-7b94-4691-b1cd-f611008439bb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.490350 4675 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f77b2245-7b94-4691-b1cd-f611008439bb-host\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.494726 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f77b2245-7b94-4691-b1cd-f611008439bb-kube-api-access-dfmd4" (OuterVolumeSpecName: "kube-api-access-dfmd4") pod "f77b2245-7b94-4691-b1cd-f611008439bb" (UID: "f77b2245-7b94-4691-b1cd-f611008439bb"). InnerVolumeSpecName "kube-api-access-dfmd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:38:09 crc kubenswrapper[4675]: I0219 09:38:09.592183 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfmd4\" (UniqueName: \"kubernetes.io/projected/f77b2245-7b94-4691-b1cd-f611008439bb-kube-api-access-dfmd4\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.230540 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9fd17ab21e62652f9bb3ef164d5b93868322b7281b3ebc807d8d70aece2c2a0" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.230669 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-vcd2j" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.609234 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-l245p/crc-debug-hnf6p"] Feb 19 09:38:10 crc kubenswrapper[4675]: E0219 09:38:10.610022 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f77b2245-7b94-4691-b1cd-f611008439bb" containerName="container-00" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.610038 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="f77b2245-7b94-4691-b1cd-f611008439bb" containerName="container-00" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.610230 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="f77b2245-7b94-4691-b1cd-f611008439bb" containerName="container-00" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.610870 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.613729 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-l245p"/"default-dockercfg-n66vl" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.712201 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbg6c\" (UniqueName: \"kubernetes.io/projected/7f46708f-4f2f-42c2-8214-fd5c068aa937-kube-api-access-dbg6c\") pod \"crc-debug-hnf6p\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.712601 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f46708f-4f2f-42c2-8214-fd5c068aa937-host\") pod \"crc-debug-hnf6p\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.814083 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f46708f-4f2f-42c2-8214-fd5c068aa937-host\") pod \"crc-debug-hnf6p\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.814181 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbg6c\" (UniqueName: \"kubernetes.io/projected/7f46708f-4f2f-42c2-8214-fd5c068aa937-kube-api-access-dbg6c\") pod \"crc-debug-hnf6p\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.814203 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f46708f-4f2f-42c2-8214-fd5c068aa937-host\") pod \"crc-debug-hnf6p\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.830548 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbg6c\" (UniqueName: \"kubernetes.io/projected/7f46708f-4f2f-42c2-8214-fd5c068aa937-kube-api-access-dbg6c\") pod \"crc-debug-hnf6p\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:10 crc kubenswrapper[4675]: I0219 09:38:10.929994 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:11 crc kubenswrapper[4675]: I0219 09:38:11.133973 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f77b2245-7b94-4691-b1cd-f611008439bb" path="/var/lib/kubelet/pods/f77b2245-7b94-4691-b1cd-f611008439bb/volumes" Feb 19 09:38:11 crc kubenswrapper[4675]: I0219 09:38:11.240426 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-hnf6p" event={"ID":"7f46708f-4f2f-42c2-8214-fd5c068aa937","Type":"ContainerStarted","Data":"b0f9f5381fe3352da749c6bb4a1c9fc54768bdd74c32d6064116b9b82e7d6b9d"} Feb 19 09:38:11 crc kubenswrapper[4675]: I0219 09:38:11.240478 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-hnf6p" event={"ID":"7f46708f-4f2f-42c2-8214-fd5c068aa937","Type":"ContainerStarted","Data":"660264f99fdbbd0e79d7454320403205824309f66bbeeb9a16a73216ba5ba4b1"} Feb 19 09:38:11 crc kubenswrapper[4675]: I0219 09:38:11.254490 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-l245p/crc-debug-hnf6p" podStartSLOduration=1.254470311 podStartE2EDuration="1.254470311s" podCreationTimestamp="2026-02-19 09:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:38:11.251134661 +0000 UTC m=+3272.878224929" watchObservedRunningTime="2026-02-19 09:38:11.254470311 +0000 UTC m=+3272.881560579" Feb 19 09:38:12 crc kubenswrapper[4675]: I0219 09:38:12.251444 4675 generic.go:334] "Generic (PLEG): container finished" podID="7f46708f-4f2f-42c2-8214-fd5c068aa937" containerID="b0f9f5381fe3352da749c6bb4a1c9fc54768bdd74c32d6064116b9b82e7d6b9d" exitCode=0 Feb 19 09:38:12 crc kubenswrapper[4675]: I0219 09:38:12.251556 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-hnf6p" event={"ID":"7f46708f-4f2f-42c2-8214-fd5c068aa937","Type":"ContainerDied","Data":"b0f9f5381fe3352da749c6bb4a1c9fc54768bdd74c32d6064116b9b82e7d6b9d"} Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.393340 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.428282 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-l245p/crc-debug-hnf6p"] Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.439987 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-l245p/crc-debug-hnf6p"] Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.465239 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbg6c\" (UniqueName: \"kubernetes.io/projected/7f46708f-4f2f-42c2-8214-fd5c068aa937-kube-api-access-dbg6c\") pod \"7f46708f-4f2f-42c2-8214-fd5c068aa937\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.465656 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f46708f-4f2f-42c2-8214-fd5c068aa937-host\") pod \"7f46708f-4f2f-42c2-8214-fd5c068aa937\" (UID: \"7f46708f-4f2f-42c2-8214-fd5c068aa937\") " Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.465791 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f46708f-4f2f-42c2-8214-fd5c068aa937-host" (OuterVolumeSpecName: "host") pod "7f46708f-4f2f-42c2-8214-fd5c068aa937" (UID: "7f46708f-4f2f-42c2-8214-fd5c068aa937"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.466311 4675 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f46708f-4f2f-42c2-8214-fd5c068aa937-host\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.471399 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f46708f-4f2f-42c2-8214-fd5c068aa937-kube-api-access-dbg6c" (OuterVolumeSpecName: "kube-api-access-dbg6c") pod "7f46708f-4f2f-42c2-8214-fd5c068aa937" (UID: "7f46708f-4f2f-42c2-8214-fd5c068aa937"). InnerVolumeSpecName "kube-api-access-dbg6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:38:13 crc kubenswrapper[4675]: I0219 09:38:13.568170 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbg6c\" (UniqueName: \"kubernetes.io/projected/7f46708f-4f2f-42c2-8214-fd5c068aa937-kube-api-access-dbg6c\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.276146 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="660264f99fdbbd0e79d7454320403205824309f66bbeeb9a16a73216ba5ba4b1" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.276249 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-hnf6p" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.571325 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-l245p/crc-debug-8krbd"] Feb 19 09:38:14 crc kubenswrapper[4675]: E0219 09:38:14.571714 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f46708f-4f2f-42c2-8214-fd5c068aa937" containerName="container-00" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.571728 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f46708f-4f2f-42c2-8214-fd5c068aa937" containerName="container-00" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.571906 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f46708f-4f2f-42c2-8214-fd5c068aa937" containerName="container-00" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.572459 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.573951 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-l245p"/"default-dockercfg-n66vl" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.691756 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09324bd4-9ed5-4ba5-843d-bff7d5f85171-host\") pod \"crc-debug-8krbd\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.692063 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnjml\" (UniqueName: \"kubernetes.io/projected/09324bd4-9ed5-4ba5-843d-bff7d5f85171-kube-api-access-dnjml\") pod \"crc-debug-8krbd\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.794269 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09324bd4-9ed5-4ba5-843d-bff7d5f85171-host\") pod \"crc-debug-8krbd\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.794421 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnjml\" (UniqueName: \"kubernetes.io/projected/09324bd4-9ed5-4ba5-843d-bff7d5f85171-kube-api-access-dnjml\") pod \"crc-debug-8krbd\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.794454 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09324bd4-9ed5-4ba5-843d-bff7d5f85171-host\") pod \"crc-debug-8krbd\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.824741 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnjml\" (UniqueName: \"kubernetes.io/projected/09324bd4-9ed5-4ba5-843d-bff7d5f85171-kube-api-access-dnjml\") pod \"crc-debug-8krbd\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: I0219 09:38:14.892851 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:14 crc kubenswrapper[4675]: W0219 09:38:14.941024 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09324bd4_9ed5_4ba5_843d_bff7d5f85171.slice/crio-eace015335e2daa1699e5f6c38bfc3d2ce2822326de3e88126f38e639f0f4987 WatchSource:0}: Error finding container eace015335e2daa1699e5f6c38bfc3d2ce2822326de3e88126f38e639f0f4987: Status 404 returned error can't find the container with id eace015335e2daa1699e5f6c38bfc3d2ce2822326de3e88126f38e639f0f4987 Feb 19 09:38:15 crc kubenswrapper[4675]: I0219 09:38:15.121082 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f46708f-4f2f-42c2-8214-fd5c068aa937" path="/var/lib/kubelet/pods/7f46708f-4f2f-42c2-8214-fd5c068aa937/volumes" Feb 19 09:38:15 crc kubenswrapper[4675]: I0219 09:38:15.314213 4675 generic.go:334] "Generic (PLEG): container finished" podID="09324bd4-9ed5-4ba5-843d-bff7d5f85171" containerID="529b27f825f97d4b12e405c7cc867017a9e8862bbeadae556d7a5287bf9121ba" exitCode=0 Feb 19 09:38:15 crc kubenswrapper[4675]: I0219 09:38:15.314280 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-8krbd" event={"ID":"09324bd4-9ed5-4ba5-843d-bff7d5f85171","Type":"ContainerDied","Data":"529b27f825f97d4b12e405c7cc867017a9e8862bbeadae556d7a5287bf9121ba"} Feb 19 09:38:15 crc kubenswrapper[4675]: I0219 09:38:15.314335 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/crc-debug-8krbd" event={"ID":"09324bd4-9ed5-4ba5-843d-bff7d5f85171","Type":"ContainerStarted","Data":"eace015335e2daa1699e5f6c38bfc3d2ce2822326de3e88126f38e639f0f4987"} Feb 19 09:38:15 crc kubenswrapper[4675]: I0219 09:38:15.363457 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-l245p/crc-debug-8krbd"] Feb 19 09:38:15 crc kubenswrapper[4675]: I0219 09:38:15.380929 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-l245p/crc-debug-8krbd"] Feb 19 09:38:16 crc kubenswrapper[4675]: I0219 09:38:16.435620 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:16 crc kubenswrapper[4675]: I0219 09:38:16.533399 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnjml\" (UniqueName: \"kubernetes.io/projected/09324bd4-9ed5-4ba5-843d-bff7d5f85171-kube-api-access-dnjml\") pod \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " Feb 19 09:38:16 crc kubenswrapper[4675]: I0219 09:38:16.533440 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09324bd4-9ed5-4ba5-843d-bff7d5f85171-host\") pod \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\" (UID: \"09324bd4-9ed5-4ba5-843d-bff7d5f85171\") " Feb 19 09:38:16 crc kubenswrapper[4675]: I0219 09:38:16.533606 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09324bd4-9ed5-4ba5-843d-bff7d5f85171-host" (OuterVolumeSpecName: "host") pod "09324bd4-9ed5-4ba5-843d-bff7d5f85171" (UID: "09324bd4-9ed5-4ba5-843d-bff7d5f85171"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:38:16 crc kubenswrapper[4675]: I0219 09:38:16.533834 4675 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09324bd4-9ed5-4ba5-843d-bff7d5f85171-host\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:16 crc kubenswrapper[4675]: I0219 09:38:16.539980 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09324bd4-9ed5-4ba5-843d-bff7d5f85171-kube-api-access-dnjml" (OuterVolumeSpecName: "kube-api-access-dnjml") pod "09324bd4-9ed5-4ba5-843d-bff7d5f85171" (UID: "09324bd4-9ed5-4ba5-843d-bff7d5f85171"). InnerVolumeSpecName "kube-api-access-dnjml". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:38:16 crc kubenswrapper[4675]: I0219 09:38:16.635457 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnjml\" (UniqueName: \"kubernetes.io/projected/09324bd4-9ed5-4ba5-843d-bff7d5f85171-kube-api-access-dnjml\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:17 crc kubenswrapper[4675]: I0219 09:38:17.116508 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09324bd4-9ed5-4ba5-843d-bff7d5f85171" path="/var/lib/kubelet/pods/09324bd4-9ed5-4ba5-843d-bff7d5f85171/volumes" Feb 19 09:38:17 crc kubenswrapper[4675]: I0219 09:38:17.337815 4675 scope.go:117] "RemoveContainer" containerID="529b27f825f97d4b12e405c7cc867017a9e8862bbeadae556d7a5287bf9121ba" Feb 19 09:38:17 crc kubenswrapper[4675]: I0219 09:38:17.337863 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/crc-debug-8krbd" Feb 19 09:38:19 crc kubenswrapper[4675]: I0219 09:38:19.110325 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:38:19 crc kubenswrapper[4675]: E0219 09:38:19.111020 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.103805 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:38:30 crc kubenswrapper[4675]: E0219 09:38:30.104489 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.416938 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b4d586cf4-9nm5g_c045dff4-aad5-44b8-a942-20a45f927859/barbican-api/0.log" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.547866 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b4d586cf4-9nm5g_c045dff4-aad5-44b8-a942-20a45f927859/barbican-api-log/0.log" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.620290 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-bcb894f54-l299f_8649bd53-643b-466d-a528-45d2aa799316/barbican-keystone-listener/0.log" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.649567 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-bcb894f54-l299f_8649bd53-643b-466d-a528-45d2aa799316/barbican-keystone-listener-log/0.log" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.819739 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7667dd7745-49ztz_5a4a2232-762f-4fc1-83fd-45d33b408e5b/barbican-worker/0.log" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.867115 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7667dd7745-49ztz_5a4a2232-762f-4fc1-83fd-45d33b408e5b/barbican-worker-log/0.log" Feb 19 09:38:30 crc kubenswrapper[4675]: I0219 09:38:30.986222 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd_df5fab89-7323-474a-96b4-60dbd70e2c50/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.079959 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/ceilometer-central-agent/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.146148 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/ceilometer-notification-agent/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.200965 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/proxy-httpd/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.252426 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/sg-core/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.362771 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b83eea9-3ec2-44bd-8498-5c5604e0eac7/cinder-api/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.409573 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b83eea9-3ec2-44bd-8498-5c5604e0eac7/cinder-api-log/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.535824 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e85d2864-3bca-456e-a6a2-b5f4bc385b15/cinder-scheduler/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.558188 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e85d2864-3bca-456e-a6a2-b5f4bc385b15/probe/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.658823 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k_8fa31ca4-5356-4a67-ac53-6bc4d9f0780f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.741777 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z_380f6f00-e968-4289-8ea9-ed1c8ef7ff59/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:31 crc kubenswrapper[4675]: I0219 09:38:31.873680 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-8grnp_012e6c07-a778-4850-99e0-47346310663b/init/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.028970 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-8grnp_012e6c07-a778-4850-99e0-47346310663b/init/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.030350 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-8grnp_012e6c07-a778-4850-99e0-47346310663b/dnsmasq-dns/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.097907 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-fr44t_31e7b53a-04a2-4fd3-b18e-893da088a60f/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.269800 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0be92ca9-2a5c-4c33-b7e4-a440f157cd02/glance-httpd/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.280074 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0be92ca9-2a5c-4c33-b7e4-a440f157cd02/glance-log/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.410407 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91a2c1f8-322b-40e7-a577-ed775f9a1565/glance-httpd/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.469373 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91a2c1f8-322b-40e7-a577-ed775f9a1565/glance-log/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.624580 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64d55559f4-zwfgc_18d90d12-7724-438d-8cd4-9fbe70c8ee89/horizon/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.742240 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-5cts8_8dad42cd-cf8e-40d0-ab92-37c57c05d7fa/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.866835 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64d55559f4-zwfgc_18d90d12-7724-438d-8cd4-9fbe70c8ee89/horizon-log/0.log" Feb 19 09:38:32 crc kubenswrapper[4675]: I0219 09:38:32.928142 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g5kn7_55d678f4-667d-4d16-8f40-4faeab6d6c1a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:33 crc kubenswrapper[4675]: I0219 09:38:33.184140 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1e70d121-b4ec-4c18-be34-18d2eb7d529e/kube-state-metrics/0.log" Feb 19 09:38:33 crc kubenswrapper[4675]: I0219 09:38:33.209901 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-8594567c89-8xr7n_a6ae7ce1-9da5-49f4-8e32-0f1e0996080a/keystone-api/0.log" Feb 19 09:38:33 crc kubenswrapper[4675]: I0219 09:38:33.331826 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pknwk_12901b5c-afde-4017-86ba-363d50c78a9a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:33 crc kubenswrapper[4675]: I0219 09:38:33.662353 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57f4c8679f-grtkm_81ba53aa-e359-412c-abe4-b91662a9996a/neutron-httpd/0.log" Feb 19 09:38:33 crc kubenswrapper[4675]: I0219 09:38:33.720233 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57f4c8679f-grtkm_81ba53aa-e359-412c-abe4-b91662a9996a/neutron-api/0.log" Feb 19 09:38:33 crc kubenswrapper[4675]: I0219 09:38:33.907794 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9_185f89ae-500c-4bdf-81b1-e3627ab1d600/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:34 crc kubenswrapper[4675]: I0219 09:38:34.407793 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_657eb3f7-7b32-44db-b26b-c070b619cee7/nova-api-log/0.log" Feb 19 09:38:34 crc kubenswrapper[4675]: I0219 09:38:34.487742 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_657eb3f7-7b32-44db-b26b-c070b619cee7/nova-api-api/0.log" Feb 19 09:38:34 crc kubenswrapper[4675]: I0219 09:38:34.533847 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_408c0ffb-6a67-4351-ac43-8c5affe01d69/nova-cell0-conductor-conductor/0.log" Feb 19 09:38:34 crc kubenswrapper[4675]: I0219 09:38:34.715123 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_670df879-c26a-46f0-b385-5bb153734e18/nova-cell1-conductor-conductor/0.log" Feb 19 09:38:34 crc kubenswrapper[4675]: I0219 09:38:34.827466 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_774eaf0e-fab7-4e49-b3de-b9e8ca212f66/nova-cell1-novncproxy-novncproxy/0.log" Feb 19 09:38:34 crc kubenswrapper[4675]: I0219 09:38:34.951700 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-r2x4h_9cee21fc-60b4-413c-b6b4-e75e813a8e9d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:35 crc kubenswrapper[4675]: I0219 09:38:35.098999 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78796291-e993-4cea-90df-f44081e71bf9/nova-metadata-log/0.log" Feb 19 09:38:35 crc kubenswrapper[4675]: I0219 09:38:35.349474 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_de054e7d-f903-4b96-8394-7d12f6cfdb98/nova-scheduler-scheduler/0.log" Feb 19 09:38:35 crc kubenswrapper[4675]: I0219 09:38:35.403920 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3b301f22-5f49-490c-899a-f518d8da2174/mysql-bootstrap/0.log" Feb 19 09:38:35 crc kubenswrapper[4675]: I0219 09:38:35.538124 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3b301f22-5f49-490c-899a-f518d8da2174/mysql-bootstrap/0.log" Feb 19 09:38:35 crc kubenswrapper[4675]: I0219 09:38:35.626299 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3b301f22-5f49-490c-899a-f518d8da2174/galera/0.log" Feb 19 09:38:35 crc kubenswrapper[4675]: I0219 09:38:35.791272 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b91a4d30-10a1-4827-b4a7-2c620a2c2d9f/mysql-bootstrap/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.065533 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78796291-e993-4cea-90df-f44081e71bf9/nova-metadata-metadata/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.068413 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b91a4d30-10a1-4827-b4a7-2c620a2c2d9f/mysql-bootstrap/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.078580 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b91a4d30-10a1-4827-b4a7-2c620a2c2d9f/galera/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.189361 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2b0f27e5-af96-47cd-9550-dfa37033f584/openstackclient/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.318410 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lnndq_10a37557-1329-4995-86a8-4dc2e12f217c/openstack-network-exporter/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.410357 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovsdb-server-init/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.558097 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovs-vswitchd/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.581238 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovsdb-server-init/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.599900 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovsdb-server/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.758520 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vlnhc_0b11fc83-5e56-43d3-a340-80686a5bdf35/ovn-controller/0.log" Feb 19 09:38:36 crc kubenswrapper[4675]: I0219 09:38:36.889805 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cghvt_5a7be99c-546f-4174-80ca-d72818c3ee43/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.027915 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d9d70128-744b-4075-ab12-f60f6c1838e7/openstack-network-exporter/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.064222 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d9d70128-744b-4075-ab12-f60f6c1838e7/ovn-northd/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.224198 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_482efc7c-8e44-4c25-903e-5c8b2cc58d0b/openstack-network-exporter/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.245166 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_482efc7c-8e44-4c25-903e-5c8b2cc58d0b/ovsdbserver-nb/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.371240 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rsk8s"] Feb 19 09:38:37 crc kubenswrapper[4675]: E0219 09:38:37.371750 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09324bd4-9ed5-4ba5-843d-bff7d5f85171" containerName="container-00" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.371774 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="09324bd4-9ed5-4ba5-843d-bff7d5f85171" containerName="container-00" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.372036 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="09324bd4-9ed5-4ba5-843d-bff7d5f85171" containerName="container-00" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.373747 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.391201 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsk8s"] Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.427549 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2344b53c-5df6-4673-beff-7c3cdd28dc93/openstack-network-exporter/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.498046 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-catalog-content\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.498157 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hjbk\" (UniqueName: \"kubernetes.io/projected/30656648-92b5-471f-81b6-680f06e02b99-kube-api-access-7hjbk\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.498186 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-utilities\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.546533 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2344b53c-5df6-4673-beff-7c3cdd28dc93/ovsdbserver-sb/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.599901 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-catalog-content\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.600009 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hjbk\" (UniqueName: \"kubernetes.io/projected/30656648-92b5-471f-81b6-680f06e02b99-kube-api-access-7hjbk\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.600042 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-utilities\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.600488 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-utilities\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.600715 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-catalog-content\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.621916 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hjbk\" (UniqueName: \"kubernetes.io/projected/30656648-92b5-471f-81b6-680f06e02b99-kube-api-access-7hjbk\") pod \"community-operators-rsk8s\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.630414 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7c56db696b-nwkrv_271c0509-6c25-44bf-8565-5d25488f67c7/placement-api/0.log" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.696797 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:37 crc kubenswrapper[4675]: I0219 09:38:37.833765 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7c56db696b-nwkrv_271c0509-6c25-44bf-8565-5d25488f67c7/placement-log/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.003101 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4/setup-container/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.148257 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4/setup-container/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.180271 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4/rabbitmq/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.277067 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsk8s"] Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.337251 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_60253071-539f-43bb-a335-b351dbac4b77/setup-container/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.516403 4675 generic.go:334] "Generic (PLEG): container finished" podID="30656648-92b5-471f-81b6-680f06e02b99" containerID="4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8" exitCode=0 Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.516730 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsk8s" event={"ID":"30656648-92b5-471f-81b6-680f06e02b99","Type":"ContainerDied","Data":"4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8"} Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.516760 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsk8s" event={"ID":"30656648-92b5-471f-81b6-680f06e02b99","Type":"ContainerStarted","Data":"ae2e0a988d59aa09474ad2904ada5bad740ee619c75e77691063558ce9a13986"} Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.585130 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_60253071-539f-43bb-a335-b351dbac4b77/setup-container/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.597000 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt_50eaff9d-656b-4e39-b450-bcc31570bc29/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.659970 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_60253071-539f-43bb-a335-b351dbac4b77/rabbitmq/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.775254 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zkwd8_185defec-09d9-4992-bc76-fc2cc62a788e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:38 crc kubenswrapper[4675]: I0219 09:38:38.866316 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp_34f8029c-b96e-42c2-8005-ebc9988389b7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.081032 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-692m6_f3bba895-fd3d-44bf-9ea7-e631560bbc0a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.154675 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-w7zd2_36e6ad34-b4ec-4e70-895b-37b09f4508aa/ssh-known-hosts-edpm-deployment/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.532902 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-676ddcfdc-554zf_c331033a-dd3e-4871-8e7d-85081578aace/proxy-server/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.533586 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsk8s" event={"ID":"30656648-92b5-471f-81b6-680f06e02b99","Type":"ContainerStarted","Data":"485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed"} Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.677163 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-676ddcfdc-554zf_c331033a-dd3e-4871-8e7d-85081578aace/proxy-httpd/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.716463 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-s4x6m_dadf56d2-1dd1-4590-8fd7-0cc266af5377/swift-ring-rebalance/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.768482 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-auditor/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.940008 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-server/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.970143 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-reaper/0.log" Feb 19 09:38:39 crc kubenswrapper[4675]: I0219 09:38:39.982481 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-auditor/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.019970 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-replicator/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.163369 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-server/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.278016 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-replicator/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.279259 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-updater/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.289958 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-auditor/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.397555 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-expirer/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.447190 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-server/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.474443 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-updater/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.485019 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-replicator/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.544799 4675 generic.go:334] "Generic (PLEG): container finished" podID="30656648-92b5-471f-81b6-680f06e02b99" containerID="485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed" exitCode=0 Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.544856 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsk8s" event={"ID":"30656648-92b5-471f-81b6-680f06e02b99","Type":"ContainerDied","Data":"485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed"} Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.615973 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/rsync/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.630396 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/swift-recon-cron/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.800823 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl_c1b6f896-16c6-40af-ae5f-4392213dacd3/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:40 crc kubenswrapper[4675]: I0219 09:38:40.917063 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5006db86-ae9d-4dbc-871f-4bc18f3a43f6/tempest-tests-tempest-tests-runner/0.log" Feb 19 09:38:41 crc kubenswrapper[4675]: I0219 09:38:41.025150 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b255b445-e2d1-4d12-8a53-34e76f18dede/test-operator-logs-container/0.log" Feb 19 09:38:41 crc kubenswrapper[4675]: I0219 09:38:41.141997 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj_af6f5807-47f8-4e48-b8c0-b04eeb4bb160/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:38:41 crc kubenswrapper[4675]: I0219 09:38:41.555948 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsk8s" event={"ID":"30656648-92b5-471f-81b6-680f06e02b99","Type":"ContainerStarted","Data":"798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b"} Feb 19 09:38:41 crc kubenswrapper[4675]: I0219 09:38:41.576441 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rsk8s" podStartSLOduration=2.080850497 podStartE2EDuration="4.576424601s" podCreationTimestamp="2026-02-19 09:38:37 +0000 UTC" firstStartedPulling="2026-02-19 09:38:38.518366147 +0000 UTC m=+3300.145456415" lastFinishedPulling="2026-02-19 09:38:41.013940251 +0000 UTC m=+3302.641030519" observedRunningTime="2026-02-19 09:38:41.573502692 +0000 UTC m=+3303.200592960" watchObservedRunningTime="2026-02-19 09:38:41.576424601 +0000 UTC m=+3303.203514869" Feb 19 09:38:42 crc kubenswrapper[4675]: I0219 09:38:42.102959 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:38:42 crc kubenswrapper[4675]: E0219 09:38:42.103229 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:38:47 crc kubenswrapper[4675]: I0219 09:38:47.697951 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:47 crc kubenswrapper[4675]: I0219 09:38:47.698468 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:47 crc kubenswrapper[4675]: I0219 09:38:47.748975 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:48 crc kubenswrapper[4675]: I0219 09:38:48.730420 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:48 crc kubenswrapper[4675]: I0219 09:38:48.900385 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsk8s"] Feb 19 09:38:48 crc kubenswrapper[4675]: I0219 09:38:48.933127 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4f51af02-6352-4462-b00f-a2feafe19f23/memcached/0.log" Feb 19 09:38:50 crc kubenswrapper[4675]: I0219 09:38:50.679072 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rsk8s" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="registry-server" containerID="cri-o://798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b" gracePeriod=2 Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.157098 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.249827 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-catalog-content\") pod \"30656648-92b5-471f-81b6-680f06e02b99\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.250205 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hjbk\" (UniqueName: \"kubernetes.io/projected/30656648-92b5-471f-81b6-680f06e02b99-kube-api-access-7hjbk\") pod \"30656648-92b5-471f-81b6-680f06e02b99\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.250303 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-utilities\") pod \"30656648-92b5-471f-81b6-680f06e02b99\" (UID: \"30656648-92b5-471f-81b6-680f06e02b99\") " Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.251186 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-utilities" (OuterVolumeSpecName: "utilities") pod "30656648-92b5-471f-81b6-680f06e02b99" (UID: "30656648-92b5-471f-81b6-680f06e02b99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.257980 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30656648-92b5-471f-81b6-680f06e02b99-kube-api-access-7hjbk" (OuterVolumeSpecName: "kube-api-access-7hjbk") pod "30656648-92b5-471f-81b6-680f06e02b99" (UID: "30656648-92b5-471f-81b6-680f06e02b99"). InnerVolumeSpecName "kube-api-access-7hjbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.304418 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30656648-92b5-471f-81b6-680f06e02b99" (UID: "30656648-92b5-471f-81b6-680f06e02b99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.352229 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.352264 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hjbk\" (UniqueName: \"kubernetes.io/projected/30656648-92b5-471f-81b6-680f06e02b99-kube-api-access-7hjbk\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.352277 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656648-92b5-471f-81b6-680f06e02b99-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.688678 4675 generic.go:334] "Generic (PLEG): container finished" podID="30656648-92b5-471f-81b6-680f06e02b99" containerID="798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b" exitCode=0 Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.688733 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsk8s" event={"ID":"30656648-92b5-471f-81b6-680f06e02b99","Type":"ContainerDied","Data":"798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b"} Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.688758 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsk8s" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.688774 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsk8s" event={"ID":"30656648-92b5-471f-81b6-680f06e02b99","Type":"ContainerDied","Data":"ae2e0a988d59aa09474ad2904ada5bad740ee619c75e77691063558ce9a13986"} Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.688800 4675 scope.go:117] "RemoveContainer" containerID="798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.710944 4675 scope.go:117] "RemoveContainer" containerID="485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.717799 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsk8s"] Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.731946 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rsk8s"] Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.739449 4675 scope.go:117] "RemoveContainer" containerID="4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.767325 4675 scope.go:117] "RemoveContainer" containerID="798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b" Feb 19 09:38:51 crc kubenswrapper[4675]: E0219 09:38:51.767891 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b\": container with ID starting with 798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b not found: ID does not exist" containerID="798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.767929 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b"} err="failed to get container status \"798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b\": rpc error: code = NotFound desc = could not find container \"798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b\": container with ID starting with 798b93e58584ff95ee05349f58cfa6c94b4ece2527c7d67db673008efc0d4c9b not found: ID does not exist" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.767955 4675 scope.go:117] "RemoveContainer" containerID="485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed" Feb 19 09:38:51 crc kubenswrapper[4675]: E0219 09:38:51.768248 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed\": container with ID starting with 485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed not found: ID does not exist" containerID="485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.768267 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed"} err="failed to get container status \"485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed\": rpc error: code = NotFound desc = could not find container \"485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed\": container with ID starting with 485a5ed4c921edba2ffd1f0445120154346765ca446f105a61acc1a5e474b2ed not found: ID does not exist" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.768280 4675 scope.go:117] "RemoveContainer" containerID="4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8" Feb 19 09:38:51 crc kubenswrapper[4675]: E0219 09:38:51.768562 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8\": container with ID starting with 4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8 not found: ID does not exist" containerID="4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8" Feb 19 09:38:51 crc kubenswrapper[4675]: I0219 09:38:51.768601 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8"} err="failed to get container status \"4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8\": rpc error: code = NotFound desc = could not find container \"4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8\": container with ID starting with 4aec71e2cc003cd355d28cf264c30d81f057d3801f5f1678862e890cda85b0d8 not found: ID does not exist" Feb 19 09:38:53 crc kubenswrapper[4675]: I0219 09:38:53.116046 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30656648-92b5-471f-81b6-680f06e02b99" path="/var/lib/kubelet/pods/30656648-92b5-471f-81b6-680f06e02b99/volumes" Feb 19 09:38:54 crc kubenswrapper[4675]: I0219 09:38:54.103989 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:38:54 crc kubenswrapper[4675]: E0219 09:38:54.104236 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:39:06 crc kubenswrapper[4675]: I0219 09:39:06.313765 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/util/0.log" Feb 19 09:39:06 crc kubenswrapper[4675]: I0219 09:39:06.445438 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/util/0.log" Feb 19 09:39:06 crc kubenswrapper[4675]: I0219 09:39:06.471851 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/pull/0.log" Feb 19 09:39:06 crc kubenswrapper[4675]: I0219 09:39:06.524516 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/pull/0.log" Feb 19 09:39:06 crc kubenswrapper[4675]: I0219 09:39:06.669478 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/extract/0.log" Feb 19 09:39:06 crc kubenswrapper[4675]: I0219 09:39:06.669593 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/pull/0.log" Feb 19 09:39:06 crc kubenswrapper[4675]: I0219 09:39:06.673075 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/util/0.log" Feb 19 09:39:07 crc kubenswrapper[4675]: I0219 09:39:07.102884 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:39:07 crc kubenswrapper[4675]: E0219 09:39:07.103195 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:39:07 crc kubenswrapper[4675]: I0219 09:39:07.126501 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-glh49_7f7fd1c2-d277-402d-a762-0ebfddc48226/manager/0.log" Feb 19 09:39:07 crc kubenswrapper[4675]: I0219 09:39:07.412485 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-v7lcl_b0de8493-3f97-4e5e-a28d-5d45f545d645/manager/0.log" Feb 19 09:39:07 crc kubenswrapper[4675]: I0219 09:39:07.623254 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-bwwwp_53318e37-b20a-46eb-8f12-decb087ade91/manager/0.log" Feb 19 09:39:07 crc kubenswrapper[4675]: I0219 09:39:07.855220 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-9zn82_2473d2e8-7c4d-41f5-bd2a-95823fe059f1/manager/0.log" Feb 19 09:39:08 crc kubenswrapper[4675]: I0219 09:39:08.099080 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-c775s_f79e0726-e5c1-4536-b06b-40bab3849bb0/manager/0.log" Feb 19 09:39:08 crc kubenswrapper[4675]: I0219 09:39:08.310926 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-z2wq2_78209850-b313-4190-a3a4-674da1afaaba/manager/0.log" Feb 19 09:39:08 crc kubenswrapper[4675]: I0219 09:39:08.426033 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-cbbkp_a5bfbb7f-0d4a-49d3-bf77-b168ccf33061/manager/0.log" Feb 19 09:39:08 crc kubenswrapper[4675]: I0219 09:39:08.550884 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-jhgw4_4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6/manager/0.log" Feb 19 09:39:08 crc kubenswrapper[4675]: I0219 09:39:08.946762 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-kqz2s_63fcb9d8-c313-4d0e-9f53-44c3ad272d1d/manager/0.log" Feb 19 09:39:09 crc kubenswrapper[4675]: I0219 09:39:09.079477 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-c25kr_61877abc-2f6b-40a5-aa19-b8adf5402cc6/manager/0.log" Feb 19 09:39:09 crc kubenswrapper[4675]: I0219 09:39:09.301955 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-8k6xs_5c9124b1-4ca0-481f-8474-ea45252bda76/manager/0.log" Feb 19 09:39:09 crc kubenswrapper[4675]: I0219 09:39:09.442243 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-ddhjt_4e163d5d-3666-4899-acd7-21040c1cc573/manager/0.log" Feb 19 09:39:09 crc kubenswrapper[4675]: I0219 09:39:09.856175 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8_769e3e7a-a9ea-4d9c-b60b-8643e0c275e1/manager/0.log" Feb 19 09:39:10 crc kubenswrapper[4675]: I0219 09:39:10.436715 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-767d85d5f6-7tx82_f7d62133-059d-4695-829f-ed7b9888e323/operator/0.log" Feb 19 09:39:10 crc kubenswrapper[4675]: I0219 09:39:10.534784 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-h9qft_8aa0b598-0a09-4439-9e9d-c125fb5542db/registry-server/0.log" Feb 19 09:39:10 crc kubenswrapper[4675]: I0219 09:39:10.874073 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-r8bkh_dcd57b86-afb5-4beb-ae73-830ce75a562f/manager/0.log" Feb 19 09:39:11 crc kubenswrapper[4675]: I0219 09:39:11.122694 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-dr2nj_2f4e8661-38a2-44bc-90ae-266f058758a9/manager/0.log" Feb 19 09:39:11 crc kubenswrapper[4675]: I0219 09:39:11.253184 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-ff4m9_4fad9fc9-cb88-4aff-89bb-0a2fd609f48a/manager/0.log" Feb 19 09:39:11 crc kubenswrapper[4675]: I0219 09:39:11.355195 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-sm2fn_b9dc6b16-af6b-470e-a9a7-d59245e8a22e/operator/0.log" Feb 19 09:39:11 crc kubenswrapper[4675]: I0219 09:39:11.496107 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-jz6fw_53989876-5869-4e41-a9c3-c99cea8347bd/manager/0.log" Feb 19 09:39:11 crc kubenswrapper[4675]: I0219 09:39:11.762379 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-s25g8_b1bbb05f-c65b-4ea1-ab19-34b888f49369/manager/0.log" Feb 19 09:39:11 crc kubenswrapper[4675]: I0219 09:39:11.793454 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-w8w5t_70a13bde-19ab-4c7c-af9a-15f1fe96d4e1/manager/0.log" Feb 19 09:39:11 crc kubenswrapper[4675]: I0219 09:39:11.996753 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-x6glz_1a383e65-12c2-43bf-b6b0-f61f20706d7a/manager/0.log" Feb 19 09:39:12 crc kubenswrapper[4675]: I0219 09:39:12.264597 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7d685cfdd5-zzm7g_a4559dd6-5152-40d8-94af-f2fff6c90e84/manager/0.log" Feb 19 09:39:13 crc kubenswrapper[4675]: I0219 09:39:13.723582 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-bj7kj_2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282/manager/0.log" Feb 19 09:39:22 crc kubenswrapper[4675]: I0219 09:39:22.103262 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:39:22 crc kubenswrapper[4675]: E0219 09:39:22.103764 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:39:29 crc kubenswrapper[4675]: I0219 09:39:29.697424 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-wf8zz_bbed8b9a-ef34-424b-81e0-457a38384ba0/control-plane-machine-set-operator/0.log" Feb 19 09:39:29 crc kubenswrapper[4675]: I0219 09:39:29.898799 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6tbj_7c39c460-3de5-4b4a-9c38-76ad1de9cbce/machine-api-operator/0.log" Feb 19 09:39:29 crc kubenswrapper[4675]: I0219 09:39:29.903859 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6tbj_7c39c460-3de5-4b4a-9c38-76ad1de9cbce/kube-rbac-proxy/0.log" Feb 19 09:39:35 crc kubenswrapper[4675]: I0219 09:39:35.104140 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:39:35 crc kubenswrapper[4675]: E0219 09:39:35.105200 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:39:42 crc kubenswrapper[4675]: I0219 09:39:42.223614 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-mfc4d_2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e/cert-manager-controller/0.log" Feb 19 09:39:42 crc kubenswrapper[4675]: I0219 09:39:42.311222 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-z9x24_68aab37f-51c0-4132-844e-bc47b4856f16/cert-manager-cainjector/0.log" Feb 19 09:39:42 crc kubenswrapper[4675]: I0219 09:39:42.420239 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-5sxn8_bf9ebddc-98fd-4515-8354-f1d7ae9fda4d/cert-manager-webhook/0.log" Feb 19 09:39:49 crc kubenswrapper[4675]: I0219 09:39:49.117380 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:39:49 crc kubenswrapper[4675]: E0219 09:39:49.118699 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:39:54 crc kubenswrapper[4675]: I0219 09:39:54.986984 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-f84lw_e6d09f36-de74-42bb-8b27-3ac3039ee87f/nmstate-console-plugin/0.log" Feb 19 09:39:55 crc kubenswrapper[4675]: I0219 09:39:55.118761 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-k28cx_d1523822-5fa6-446b-9cd1-3442536238f2/nmstate-handler/0.log" Feb 19 09:39:55 crc kubenswrapper[4675]: I0219 09:39:55.173584 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-92cqz_8a261a06-8401-488e-8b98-d885f1b694cf/kube-rbac-proxy/0.log" Feb 19 09:39:55 crc kubenswrapper[4675]: I0219 09:39:55.207532 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-92cqz_8a261a06-8401-488e-8b98-d885f1b694cf/nmstate-metrics/0.log" Feb 19 09:39:55 crc kubenswrapper[4675]: I0219 09:39:55.358302 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-mmtc8_a6fb0b89-c8be-4b62-a92b-23197720e978/nmstate-operator/0.log" Feb 19 09:39:55 crc kubenswrapper[4675]: I0219 09:39:55.417505 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-chcgg_3be88498-1577-460a-9431-4a4bc66eb217/nmstate-webhook/0.log" Feb 19 09:40:04 crc kubenswrapper[4675]: I0219 09:40:04.103752 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:40:04 crc kubenswrapper[4675]: E0219 09:40:04.104473 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:40:18 crc kubenswrapper[4675]: I0219 09:40:18.103477 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:40:18 crc kubenswrapper[4675]: E0219 09:40:18.104264 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:40:20 crc kubenswrapper[4675]: I0219 09:40:20.873616 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-j4tln_47ca0e9d-940d-4952-9780-74e58570d98e/kube-rbac-proxy/0.log" Feb 19 09:40:20 crc kubenswrapper[4675]: I0219 09:40:20.909356 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-j4tln_47ca0e9d-940d-4952-9780-74e58570d98e/controller/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.091804 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.239832 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.250787 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.283767 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.313435 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.494226 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.505767 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.509818 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.515641 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.658438 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.666168 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.666327 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.669610 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/controller/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.851587 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/frr-metrics/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.856569 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/kube-rbac-proxy-frr/0.log" Feb 19 09:40:21 crc kubenswrapper[4675]: I0219 09:40:21.862041 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/kube-rbac-proxy/0.log" Feb 19 09:40:22 crc kubenswrapper[4675]: I0219 09:40:22.025776 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/reloader/0.log" Feb 19 09:40:22 crc kubenswrapper[4675]: I0219 09:40:22.275176 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-fjr4t_ee8232ef-f035-4ef4-b867-98508be2d0e7/frr-k8s-webhook-server/0.log" Feb 19 09:40:22 crc kubenswrapper[4675]: I0219 09:40:22.291402 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-698bb84f65-v6wc9_8e1476c5-559f-4da9-9092-ef6b08253b6a/manager/0.log" Feb 19 09:40:22 crc kubenswrapper[4675]: I0219 09:40:22.445127 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-85dcc8c87f-62fjw_d7d966d2-c2fa-411d-8474-ab26f3cbea46/webhook-server/0.log" Feb 19 09:40:22 crc kubenswrapper[4675]: I0219 09:40:22.729255 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bc8vc_75ab6240-327c-4ac4-94f2-3d4c44e1a664/kube-rbac-proxy/0.log" Feb 19 09:40:23 crc kubenswrapper[4675]: I0219 09:40:23.152016 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bc8vc_75ab6240-327c-4ac4-94f2-3d4c44e1a664/speaker/0.log" Feb 19 09:40:23 crc kubenswrapper[4675]: I0219 09:40:23.488142 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/frr/0.log" Feb 19 09:40:33 crc kubenswrapper[4675]: I0219 09:40:33.103919 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:40:33 crc kubenswrapper[4675]: E0219 09:40:33.104727 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:40:34 crc kubenswrapper[4675]: I0219 09:40:34.832416 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/util/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.207764 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/pull/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.218535 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/util/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.278088 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/pull/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.379806 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/util/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.394206 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/pull/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.444134 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/extract/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.554873 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-utilities/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.734020 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-utilities/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.734229 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-content/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.769973 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-content/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.896860 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-content/0.log" Feb 19 09:40:35 crc kubenswrapper[4675]: I0219 09:40:35.991615 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-utilities/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.167120 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-utilities/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.317701 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/registry-server/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.340006 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-utilities/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.376743 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-content/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.438207 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-content/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.525113 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-content/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.589073 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-utilities/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.758674 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/util/0.log" Feb 19 09:40:36 crc kubenswrapper[4675]: I0219 09:40:36.990662 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/util/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.023677 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/pull/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.106403 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/pull/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.265908 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/util/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.266211 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/registry-server/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.326059 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/extract/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.331192 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/pull/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.483677 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-c4g5q_147346d6-fe78-404a-964e-4ee12c505b82/marketplace-operator/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.545344 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-utilities/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.825623 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-utilities/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.840718 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-content/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.846784 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-content/0.log" Feb 19 09:40:37 crc kubenswrapper[4675]: I0219 09:40:37.980549 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-utilities/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.055458 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-content/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.187708 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-utilities/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.251106 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/registry-server/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.419181 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-content/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.423687 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-content/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.432486 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-utilities/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.623528 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-content/0.log" Feb 19 09:40:38 crc kubenswrapper[4675]: I0219 09:40:38.623836 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-utilities/0.log" Feb 19 09:40:39 crc kubenswrapper[4675]: I0219 09:40:39.007291 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/registry-server/0.log" Feb 19 09:40:47 crc kubenswrapper[4675]: I0219 09:40:47.103434 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:40:48 crc kubenswrapper[4675]: I0219 09:40:48.063590 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"e250573cac06229729a4b58db31b629a73c1df55fa477fcc7c737ad0061e5267"} Feb 19 09:42:22 crc kubenswrapper[4675]: I0219 09:42:22.899972 4675 generic.go:334] "Generic (PLEG): container finished" podID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerID="0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae" exitCode=0 Feb 19 09:42:22 crc kubenswrapper[4675]: I0219 09:42:22.900027 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-l245p/must-gather-qz4d2" event={"ID":"0b7669be-8127-4b9b-a79d-b94ac608cb55","Type":"ContainerDied","Data":"0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae"} Feb 19 09:42:22 crc kubenswrapper[4675]: I0219 09:42:22.900898 4675 scope.go:117] "RemoveContainer" containerID="0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae" Feb 19 09:42:23 crc kubenswrapper[4675]: I0219 09:42:23.731115 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-l245p_must-gather-qz4d2_0b7669be-8127-4b9b-a79d-b94ac608cb55/gather/0.log" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.153065 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-l245p/must-gather-qz4d2"] Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.153692 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-l245p/must-gather-qz4d2" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerName="copy" containerID="cri-o://cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86" gracePeriod=2 Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.159754 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-l245p/must-gather-qz4d2"] Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.576539 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-l245p_must-gather-qz4d2_0b7669be-8127-4b9b-a79d-b94ac608cb55/copy/0.log" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.577097 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.652838 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0b7669be-8127-4b9b-a79d-b94ac608cb55-must-gather-output\") pod \"0b7669be-8127-4b9b-a79d-b94ac608cb55\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.653096 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6hqj\" (UniqueName: \"kubernetes.io/projected/0b7669be-8127-4b9b-a79d-b94ac608cb55-kube-api-access-x6hqj\") pod \"0b7669be-8127-4b9b-a79d-b94ac608cb55\" (UID: \"0b7669be-8127-4b9b-a79d-b94ac608cb55\") " Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.660128 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7669be-8127-4b9b-a79d-b94ac608cb55-kube-api-access-x6hqj" (OuterVolumeSpecName: "kube-api-access-x6hqj") pod "0b7669be-8127-4b9b-a79d-b94ac608cb55" (UID: "0b7669be-8127-4b9b-a79d-b94ac608cb55"). InnerVolumeSpecName "kube-api-access-x6hqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.754904 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6hqj\" (UniqueName: \"kubernetes.io/projected/0b7669be-8127-4b9b-a79d-b94ac608cb55-kube-api-access-x6hqj\") on node \"crc\" DevicePath \"\"" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.819135 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b7669be-8127-4b9b-a79d-b94ac608cb55-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0b7669be-8127-4b9b-a79d-b94ac608cb55" (UID: "0b7669be-8127-4b9b-a79d-b94ac608cb55"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.856606 4675 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0b7669be-8127-4b9b-a79d-b94ac608cb55-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.974737 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-l245p_must-gather-qz4d2_0b7669be-8127-4b9b-a79d-b94ac608cb55/copy/0.log" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.975170 4675 generic.go:334] "Generic (PLEG): container finished" podID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerID="cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86" exitCode=143 Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.975236 4675 scope.go:117] "RemoveContainer" containerID="cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.975237 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-l245p/must-gather-qz4d2" Feb 19 09:42:31 crc kubenswrapper[4675]: I0219 09:42:31.992048 4675 scope.go:117] "RemoveContainer" containerID="0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae" Feb 19 09:42:32 crc kubenswrapper[4675]: I0219 09:42:32.063558 4675 scope.go:117] "RemoveContainer" containerID="cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86" Feb 19 09:42:32 crc kubenswrapper[4675]: E0219 09:42:32.063948 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86\": container with ID starting with cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86 not found: ID does not exist" containerID="cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86" Feb 19 09:42:32 crc kubenswrapper[4675]: I0219 09:42:32.063980 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86"} err="failed to get container status \"cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86\": rpc error: code = NotFound desc = could not find container \"cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86\": container with ID starting with cd1be3f3ef675d7dcd6e6f6ed6e7e1bc4e584b37c053552f76b907fd59879f86 not found: ID does not exist" Feb 19 09:42:32 crc kubenswrapper[4675]: I0219 09:42:32.064000 4675 scope.go:117] "RemoveContainer" containerID="0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae" Feb 19 09:42:32 crc kubenswrapper[4675]: E0219 09:42:32.064207 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae\": container with ID starting with 0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae not found: ID does not exist" containerID="0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae" Feb 19 09:42:32 crc kubenswrapper[4675]: I0219 09:42:32.064236 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae"} err="failed to get container status \"0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae\": rpc error: code = NotFound desc = could not find container \"0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae\": container with ID starting with 0e6a799128d39abefe1b3a0f576f16fa64dfe8486516caed2f273b5037ecf6ae not found: ID does not exist" Feb 19 09:42:33 crc kubenswrapper[4675]: I0219 09:42:33.117042 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" path="/var/lib/kubelet/pods/0b7669be-8127-4b9b-a79d-b94ac608cb55/volumes" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.971200 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v4pm4"] Feb 19 09:42:42 crc kubenswrapper[4675]: E0219 09:42:42.973874 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerName="copy" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.973890 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerName="copy" Feb 19 09:42:42 crc kubenswrapper[4675]: E0219 09:42:42.973898 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="registry-server" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.973905 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="registry-server" Feb 19 09:42:42 crc kubenswrapper[4675]: E0219 09:42:42.973922 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerName="gather" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.973928 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerName="gather" Feb 19 09:42:42 crc kubenswrapper[4675]: E0219 09:42:42.973939 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="extract-utilities" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.973945 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="extract-utilities" Feb 19 09:42:42 crc kubenswrapper[4675]: E0219 09:42:42.973975 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="extract-content" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.973981 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="extract-content" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.974134 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="30656648-92b5-471f-81b6-680f06e02b99" containerName="registry-server" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.974146 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerName="copy" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.974163 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7669be-8127-4b9b-a79d-b94ac608cb55" containerName="gather" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.975474 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:42 crc kubenswrapper[4675]: I0219 09:42:42.987295 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v4pm4"] Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.058351 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-catalog-content\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.058429 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-utilities\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.058517 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpfsl\" (UniqueName: \"kubernetes.io/projected/7556566f-c32c-420b-a586-11d985fa24c5-kube-api-access-fpfsl\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.160498 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-catalog-content\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.160581 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-utilities\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.160676 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpfsl\" (UniqueName: \"kubernetes.io/projected/7556566f-c32c-420b-a586-11d985fa24c5-kube-api-access-fpfsl\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.161188 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-utilities\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.161196 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-catalog-content\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.198094 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpfsl\" (UniqueName: \"kubernetes.io/projected/7556566f-c32c-420b-a586-11d985fa24c5-kube-api-access-fpfsl\") pod \"redhat-operators-v4pm4\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.294377 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:43 crc kubenswrapper[4675]: I0219 09:42:43.792237 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v4pm4"] Feb 19 09:42:44 crc kubenswrapper[4675]: I0219 09:42:44.068040 4675 generic.go:334] "Generic (PLEG): container finished" podID="7556566f-c32c-420b-a586-11d985fa24c5" containerID="4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075" exitCode=0 Feb 19 09:42:44 crc kubenswrapper[4675]: I0219 09:42:44.068114 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v4pm4" event={"ID":"7556566f-c32c-420b-a586-11d985fa24c5","Type":"ContainerDied","Data":"4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075"} Feb 19 09:42:44 crc kubenswrapper[4675]: I0219 09:42:44.068399 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v4pm4" event={"ID":"7556566f-c32c-420b-a586-11d985fa24c5","Type":"ContainerStarted","Data":"ac66b86ae30eeca787b7cc5fb26d9959451c4f52ccdd67f8c68943f637d1cb8d"} Feb 19 09:42:44 crc kubenswrapper[4675]: I0219 09:42:44.070940 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:42:46 crc kubenswrapper[4675]: I0219 09:42:46.084415 4675 generic.go:334] "Generic (PLEG): container finished" podID="7556566f-c32c-420b-a586-11d985fa24c5" containerID="e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae" exitCode=0 Feb 19 09:42:46 crc kubenswrapper[4675]: I0219 09:42:46.084687 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v4pm4" event={"ID":"7556566f-c32c-420b-a586-11d985fa24c5","Type":"ContainerDied","Data":"e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae"} Feb 19 09:42:49 crc kubenswrapper[4675]: I0219 09:42:49.115790 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v4pm4" event={"ID":"7556566f-c32c-420b-a586-11d985fa24c5","Type":"ContainerStarted","Data":"f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136"} Feb 19 09:42:49 crc kubenswrapper[4675]: I0219 09:42:49.141149 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v4pm4" podStartSLOduration=4.69862731 podStartE2EDuration="7.141131184s" podCreationTimestamp="2026-02-19 09:42:42 +0000 UTC" firstStartedPulling="2026-02-19 09:42:44.070568283 +0000 UTC m=+3545.697658561" lastFinishedPulling="2026-02-19 09:42:46.513072167 +0000 UTC m=+3548.140162435" observedRunningTime="2026-02-19 09:42:49.134373823 +0000 UTC m=+3550.761464081" watchObservedRunningTime="2026-02-19 09:42:49.141131184 +0000 UTC m=+3550.768221452" Feb 19 09:42:53 crc kubenswrapper[4675]: I0219 09:42:53.294991 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:53 crc kubenswrapper[4675]: I0219 09:42:53.295368 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:42:54 crc kubenswrapper[4675]: I0219 09:42:54.353405 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v4pm4" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="registry-server" probeResult="failure" output=< Feb 19 09:42:54 crc kubenswrapper[4675]: timeout: failed to connect service ":50051" within 1s Feb 19 09:42:54 crc kubenswrapper[4675]: > Feb 19 09:43:03 crc kubenswrapper[4675]: I0219 09:43:03.364565 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:43:03 crc kubenswrapper[4675]: I0219 09:43:03.416821 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:43:03 crc kubenswrapper[4675]: I0219 09:43:03.610096 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v4pm4"] Feb 19 09:43:05 crc kubenswrapper[4675]: I0219 09:43:05.255428 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v4pm4" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="registry-server" containerID="cri-o://f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136" gracePeriod=2 Feb 19 09:43:05 crc kubenswrapper[4675]: I0219 09:43:05.737195 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:43:05 crc kubenswrapper[4675]: I0219 09:43:05.898738 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-catalog-content\") pod \"7556566f-c32c-420b-a586-11d985fa24c5\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " Feb 19 09:43:05 crc kubenswrapper[4675]: I0219 09:43:05.898852 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpfsl\" (UniqueName: \"kubernetes.io/projected/7556566f-c32c-420b-a586-11d985fa24c5-kube-api-access-fpfsl\") pod \"7556566f-c32c-420b-a586-11d985fa24c5\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " Feb 19 09:43:05 crc kubenswrapper[4675]: I0219 09:43:05.899000 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-utilities\") pod \"7556566f-c32c-420b-a586-11d985fa24c5\" (UID: \"7556566f-c32c-420b-a586-11d985fa24c5\") " Feb 19 09:43:05 crc kubenswrapper[4675]: I0219 09:43:05.900190 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-utilities" (OuterVolumeSpecName: "utilities") pod "7556566f-c32c-420b-a586-11d985fa24c5" (UID: "7556566f-c32c-420b-a586-11d985fa24c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:43:05 crc kubenswrapper[4675]: I0219 09:43:05.904760 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7556566f-c32c-420b-a586-11d985fa24c5-kube-api-access-fpfsl" (OuterVolumeSpecName: "kube-api-access-fpfsl") pod "7556566f-c32c-420b-a586-11d985fa24c5" (UID: "7556566f-c32c-420b-a586-11d985fa24c5"). InnerVolumeSpecName "kube-api-access-fpfsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.001261 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpfsl\" (UniqueName: \"kubernetes.io/projected/7556566f-c32c-420b-a586-11d985fa24c5-kube-api-access-fpfsl\") on node \"crc\" DevicePath \"\"" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.001300 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.041809 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7556566f-c32c-420b-a586-11d985fa24c5" (UID: "7556566f-c32c-420b-a586-11d985fa24c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.102570 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7556566f-c32c-420b-a586-11d985fa24c5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.266657 4675 generic.go:334] "Generic (PLEG): container finished" podID="7556566f-c32c-420b-a586-11d985fa24c5" containerID="f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136" exitCode=0 Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.266844 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v4pm4" event={"ID":"7556566f-c32c-420b-a586-11d985fa24c5","Type":"ContainerDied","Data":"f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136"} Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.267011 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v4pm4" event={"ID":"7556566f-c32c-420b-a586-11d985fa24c5","Type":"ContainerDied","Data":"ac66b86ae30eeca787b7cc5fb26d9959451c4f52ccdd67f8c68943f637d1cb8d"} Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.267031 4675 scope.go:117] "RemoveContainer" containerID="f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.270849 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v4pm4" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.289022 4675 scope.go:117] "RemoveContainer" containerID="e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.314342 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v4pm4"] Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.321578 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v4pm4"] Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.327741 4675 scope.go:117] "RemoveContainer" containerID="4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.391226 4675 scope.go:117] "RemoveContainer" containerID="f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136" Feb 19 09:43:06 crc kubenswrapper[4675]: E0219 09:43:06.391709 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136\": container with ID starting with f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136 not found: ID does not exist" containerID="f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.391742 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136"} err="failed to get container status \"f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136\": rpc error: code = NotFound desc = could not find container \"f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136\": container with ID starting with f98c8f47a7d5418d9408a1f1d36e0f7d81acbe464a16c8b76695b73947cfe136 not found: ID does not exist" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.391761 4675 scope.go:117] "RemoveContainer" containerID="e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae" Feb 19 09:43:06 crc kubenswrapper[4675]: E0219 09:43:06.392131 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae\": container with ID starting with e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae not found: ID does not exist" containerID="e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.392172 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae"} err="failed to get container status \"e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae\": rpc error: code = NotFound desc = could not find container \"e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae\": container with ID starting with e325082ea17e625abf00b3c1c4f54fb56f54bf3cb9f34c5a5bb8f879b47ca7ae not found: ID does not exist" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.392195 4675 scope.go:117] "RemoveContainer" containerID="4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075" Feb 19 09:43:06 crc kubenswrapper[4675]: E0219 09:43:06.393416 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075\": container with ID starting with 4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075 not found: ID does not exist" containerID="4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075" Feb 19 09:43:06 crc kubenswrapper[4675]: I0219 09:43:06.393477 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075"} err="failed to get container status \"4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075\": rpc error: code = NotFound desc = could not find container \"4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075\": container with ID starting with 4a021dc0f490de290851ddb37b0413abb44f5b97cbd7270fb52f659609dad075 not found: ID does not exist" Feb 19 09:43:07 crc kubenswrapper[4675]: I0219 09:43:07.117585 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7556566f-c32c-420b-a586-11d985fa24c5" path="/var/lib/kubelet/pods/7556566f-c32c-420b-a586-11d985fa24c5/volumes" Feb 19 09:43:11 crc kubenswrapper[4675]: I0219 09:43:11.743491 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:43:11 crc kubenswrapper[4675]: I0219 09:43:11.744176 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:43:41 crc kubenswrapper[4675]: I0219 09:43:41.743428 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:43:41 crc kubenswrapper[4675]: I0219 09:43:41.744028 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:43:59 crc kubenswrapper[4675]: I0219 09:43:59.160014 4675 scope.go:117] "RemoveContainer" containerID="86820b4f1b3cebbf94aba162f6db04d3638997282615c85e44e8ae3973e85b31" Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.743451 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.744266 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.744314 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.745031 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e250573cac06229729a4b58db31b629a73c1df55fa477fcc7c737ad0061e5267"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.745084 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://e250573cac06229729a4b58db31b629a73c1df55fa477fcc7c737ad0061e5267" gracePeriod=600 Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.920358 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="e250573cac06229729a4b58db31b629a73c1df55fa477fcc7c737ad0061e5267" exitCode=0 Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.920410 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"e250573cac06229729a4b58db31b629a73c1df55fa477fcc7c737ad0061e5267"} Feb 19 09:44:11 crc kubenswrapper[4675]: I0219 09:44:11.920447 4675 scope.go:117] "RemoveContainer" containerID="537a1bc725f79144c2d72ba78cca68041d36163ed647808990cab62d233df687" Feb 19 09:44:12 crc kubenswrapper[4675]: I0219 09:44:12.929229 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a"} Feb 19 09:44:59 crc kubenswrapper[4675]: I0219 09:44:59.249364 4675 scope.go:117] "RemoveContainer" containerID="b0f9f5381fe3352da749c6bb4a1c9fc54768bdd74c32d6064116b9b82e7d6b9d" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.171271 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw"] Feb 19 09:45:00 crc kubenswrapper[4675]: E0219 09:45:00.171886 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="extract-content" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.171898 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="extract-content" Feb 19 09:45:00 crc kubenswrapper[4675]: E0219 09:45:00.171922 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="extract-utilities" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.171928 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="extract-utilities" Feb 19 09:45:00 crc kubenswrapper[4675]: E0219 09:45:00.171949 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="registry-server" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.171958 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="registry-server" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.172141 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="7556566f-c32c-420b-a586-11d985fa24c5" containerName="registry-server" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.172734 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.175532 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.176469 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.190514 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw"] Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.269193 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-secret-volume\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.269586 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-config-volume\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.269797 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxv6b\" (UniqueName: \"kubernetes.io/projected/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-kube-api-access-vxv6b\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.371222 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-config-volume\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.371268 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxv6b\" (UniqueName: \"kubernetes.io/projected/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-kube-api-access-vxv6b\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.371361 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-secret-volume\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.372282 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-config-volume\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.380631 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-secret-volume\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.390496 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxv6b\" (UniqueName: \"kubernetes.io/projected/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-kube-api-access-vxv6b\") pod \"collect-profiles-29524905-rvxmw\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:00 crc kubenswrapper[4675]: I0219 09:45:00.549898 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:01 crc kubenswrapper[4675]: I0219 09:45:01.004514 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw"] Feb 19 09:45:01 crc kubenswrapper[4675]: I0219 09:45:01.388033 4675 generic.go:334] "Generic (PLEG): container finished" podID="ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef" containerID="05187baebd40b6b41d418dd5ee6faf4daee50f1f3e8c1a94923988a9e7c58fec" exitCode=0 Feb 19 09:45:01 crc kubenswrapper[4675]: I0219 09:45:01.388127 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" event={"ID":"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef","Type":"ContainerDied","Data":"05187baebd40b6b41d418dd5ee6faf4daee50f1f3e8c1a94923988a9e7c58fec"} Feb 19 09:45:01 crc kubenswrapper[4675]: I0219 09:45:01.388341 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" event={"ID":"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef","Type":"ContainerStarted","Data":"2f6a9e5be833b7fdb94c59d26311f85663d248e9887fce45232ab24426707b3d"} Feb 19 09:45:02 crc kubenswrapper[4675]: I0219 09:45:02.747179 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:02 crc kubenswrapper[4675]: I0219 09:45:02.914726 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-secret-volume\") pod \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " Feb 19 09:45:02 crc kubenswrapper[4675]: I0219 09:45:02.914887 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-config-volume\") pod \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " Feb 19 09:45:02 crc kubenswrapper[4675]: I0219 09:45:02.915032 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxv6b\" (UniqueName: \"kubernetes.io/projected/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-kube-api-access-vxv6b\") pod \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\" (UID: \"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef\") " Feb 19 09:45:02 crc kubenswrapper[4675]: I0219 09:45:02.916354 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-config-volume" (OuterVolumeSpecName: "config-volume") pod "ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef" (UID: "ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 19 09:45:02 crc kubenswrapper[4675]: I0219 09:45:02.922826 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef" (UID: "ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 19 09:45:02 crc kubenswrapper[4675]: I0219 09:45:02.930006 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-kube-api-access-vxv6b" (OuterVolumeSpecName: "kube-api-access-vxv6b") pod "ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef" (UID: "ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef"). InnerVolumeSpecName "kube-api-access-vxv6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.018857 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxv6b\" (UniqueName: \"kubernetes.io/projected/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-kube-api-access-vxv6b\") on node \"crc\" DevicePath \"\"" Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.019159 4675 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.019179 4675 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef-config-volume\") on node \"crc\" DevicePath \"\"" Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.405774 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" event={"ID":"ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef","Type":"ContainerDied","Data":"2f6a9e5be833b7fdb94c59d26311f85663d248e9887fce45232ab24426707b3d"} Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.405834 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29524905-rvxmw" Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.405842 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f6a9e5be833b7fdb94c59d26311f85663d248e9887fce45232ab24426707b3d" Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.825696 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8"] Feb 19 09:45:03 crc kubenswrapper[4675]: I0219 09:45:03.834127 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29524860-mf7w8"] Feb 19 09:45:05 crc kubenswrapper[4675]: I0219 09:45:05.117559 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a68599d-2cc4-4323-8437-58d0866e994e" path="/var/lib/kubelet/pods/7a68599d-2cc4-4323-8437-58d0866e994e/volumes" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.133173 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b4gs/must-gather-rxbnt"] Feb 19 09:45:28 crc kubenswrapper[4675]: E0219 09:45:28.134223 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef" containerName="collect-profiles" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.134242 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef" containerName="collect-profiles" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.134545 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca6513ce-ddaf-40ac-b368-08f3ebe1f7ef" containerName="collect-profiles" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.135800 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.139759 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7b4gs"/"kube-root-ca.crt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.140324 4675 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7b4gs"/"openshift-service-ca.crt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.143290 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7b4gs/must-gather-rxbnt"] Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.146509 4675 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7b4gs"/"default-dockercfg-xhx4n" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.192121 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nh8s\" (UniqueName: \"kubernetes.io/projected/fab9926e-edf2-4c5a-bed3-40d59f293458-kube-api-access-9nh8s\") pod \"must-gather-rxbnt\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.192181 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fab9926e-edf2-4c5a-bed3-40d59f293458-must-gather-output\") pod \"must-gather-rxbnt\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.294068 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nh8s\" (UniqueName: \"kubernetes.io/projected/fab9926e-edf2-4c5a-bed3-40d59f293458-kube-api-access-9nh8s\") pod \"must-gather-rxbnt\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.294158 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fab9926e-edf2-4c5a-bed3-40d59f293458-must-gather-output\") pod \"must-gather-rxbnt\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.294692 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fab9926e-edf2-4c5a-bed3-40d59f293458-must-gather-output\") pod \"must-gather-rxbnt\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.313414 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nh8s\" (UniqueName: \"kubernetes.io/projected/fab9926e-edf2-4c5a-bed3-40d59f293458-kube-api-access-9nh8s\") pod \"must-gather-rxbnt\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.453545 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:45:28 crc kubenswrapper[4675]: I0219 09:45:28.767448 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7b4gs/must-gather-rxbnt"] Feb 19 09:45:29 crc kubenswrapper[4675]: I0219 09:45:29.653399 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" event={"ID":"fab9926e-edf2-4c5a-bed3-40d59f293458","Type":"ContainerStarted","Data":"7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b"} Feb 19 09:45:29 crc kubenswrapper[4675]: I0219 09:45:29.653925 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" event={"ID":"fab9926e-edf2-4c5a-bed3-40d59f293458","Type":"ContainerStarted","Data":"597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc"} Feb 19 09:45:29 crc kubenswrapper[4675]: I0219 09:45:29.653946 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" event={"ID":"fab9926e-edf2-4c5a-bed3-40d59f293458","Type":"ContainerStarted","Data":"72d5bfecac47a0a1b4bcccdb855658fa24ce83a75832c0259d52af7800e47d89"} Feb 19 09:45:29 crc kubenswrapper[4675]: I0219 09:45:29.671345 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" podStartSLOduration=1.671320073 podStartE2EDuration="1.671320073s" podCreationTimestamp="2026-02-19 09:45:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:45:29.666263938 +0000 UTC m=+3711.293354206" watchObservedRunningTime="2026-02-19 09:45:29.671320073 +0000 UTC m=+3711.298410341" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.346609 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-45vsx"] Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.348218 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.375953 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpg2p\" (UniqueName: \"kubernetes.io/projected/452f7aff-16e4-41f5-89ce-ec7be419cc5a-kube-api-access-fpg2p\") pod \"crc-debug-45vsx\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.376021 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/452f7aff-16e4-41f5-89ce-ec7be419cc5a-host\") pod \"crc-debug-45vsx\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.477741 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpg2p\" (UniqueName: \"kubernetes.io/projected/452f7aff-16e4-41f5-89ce-ec7be419cc5a-kube-api-access-fpg2p\") pod \"crc-debug-45vsx\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.477816 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/452f7aff-16e4-41f5-89ce-ec7be419cc5a-host\") pod \"crc-debug-45vsx\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.477936 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/452f7aff-16e4-41f5-89ce-ec7be419cc5a-host\") pod \"crc-debug-45vsx\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.495948 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpg2p\" (UniqueName: \"kubernetes.io/projected/452f7aff-16e4-41f5-89ce-ec7be419cc5a-kube-api-access-fpg2p\") pod \"crc-debug-45vsx\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:32 crc kubenswrapper[4675]: I0219 09:45:32.666911 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:45:33 crc kubenswrapper[4675]: I0219 09:45:33.701107 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" event={"ID":"452f7aff-16e4-41f5-89ce-ec7be419cc5a","Type":"ContainerStarted","Data":"4005f0753a3bf37f95ee6c7bf9b917b9cfddf700a95228e9733838d2762d75c0"} Feb 19 09:45:33 crc kubenswrapper[4675]: I0219 09:45:33.701553 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" event={"ID":"452f7aff-16e4-41f5-89ce-ec7be419cc5a","Type":"ContainerStarted","Data":"423190c6499934b9e5ae319b788b3c34259ea3164c324fce9a5dbb56300d6663"} Feb 19 09:45:33 crc kubenswrapper[4675]: I0219 09:45:33.714953 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" podStartSLOduration=1.7149393929999999 podStartE2EDuration="1.714939393s" podCreationTimestamp="2026-02-19 09:45:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-19 09:45:33.713687799 +0000 UTC m=+3715.340778057" watchObservedRunningTime="2026-02-19 09:45:33.714939393 +0000 UTC m=+3715.342029661" Feb 19 09:45:59 crc kubenswrapper[4675]: I0219 09:45:59.299478 4675 scope.go:117] "RemoveContainer" containerID="c84abb88f7fe6735590038f2294263e3ef8bea42d52051f39cbff23483fd4029" Feb 19 09:46:04 crc kubenswrapper[4675]: I0219 09:46:04.969238 4675 generic.go:334] "Generic (PLEG): container finished" podID="452f7aff-16e4-41f5-89ce-ec7be419cc5a" containerID="4005f0753a3bf37f95ee6c7bf9b917b9cfddf700a95228e9733838d2762d75c0" exitCode=0 Feb 19 09:46:04 crc kubenswrapper[4675]: I0219 09:46:04.969800 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" event={"ID":"452f7aff-16e4-41f5-89ce-ec7be419cc5a","Type":"ContainerDied","Data":"4005f0753a3bf37f95ee6c7bf9b917b9cfddf700a95228e9733838d2762d75c0"} Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.087259 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.124607 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-45vsx"] Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.142148 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-45vsx"] Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.213050 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpg2p\" (UniqueName: \"kubernetes.io/projected/452f7aff-16e4-41f5-89ce-ec7be419cc5a-kube-api-access-fpg2p\") pod \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.213230 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/452f7aff-16e4-41f5-89ce-ec7be419cc5a-host\") pod \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\" (UID: \"452f7aff-16e4-41f5-89ce-ec7be419cc5a\") " Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.213286 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/452f7aff-16e4-41f5-89ce-ec7be419cc5a-host" (OuterVolumeSpecName: "host") pod "452f7aff-16e4-41f5-89ce-ec7be419cc5a" (UID: "452f7aff-16e4-41f5-89ce-ec7be419cc5a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.213725 4675 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/452f7aff-16e4-41f5-89ce-ec7be419cc5a-host\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.219538 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452f7aff-16e4-41f5-89ce-ec7be419cc5a-kube-api-access-fpg2p" (OuterVolumeSpecName: "kube-api-access-fpg2p") pod "452f7aff-16e4-41f5-89ce-ec7be419cc5a" (UID: "452f7aff-16e4-41f5-89ce-ec7be419cc5a"). InnerVolumeSpecName "kube-api-access-fpg2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.315971 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpg2p\" (UniqueName: \"kubernetes.io/projected/452f7aff-16e4-41f5-89ce-ec7be419cc5a-kube-api-access-fpg2p\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.985557 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="423190c6499934b9e5ae319b788b3c34259ea3164c324fce9a5dbb56300d6663" Feb 19 09:46:06 crc kubenswrapper[4675]: I0219 09:46:06.985716 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-45vsx" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.113275 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="452f7aff-16e4-41f5-89ce-ec7be419cc5a" path="/var/lib/kubelet/pods/452f7aff-16e4-41f5-89ce-ec7be419cc5a/volumes" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.381050 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-xgqwz"] Feb 19 09:46:07 crc kubenswrapper[4675]: E0219 09:46:07.381549 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452f7aff-16e4-41f5-89ce-ec7be419cc5a" containerName="container-00" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.381576 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="452f7aff-16e4-41f5-89ce-ec7be419cc5a" containerName="container-00" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.381835 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="452f7aff-16e4-41f5-89ce-ec7be419cc5a" containerName="container-00" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.382572 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.539244 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9104935c-53e2-4681-9e94-04a7354bba56-host\") pod \"crc-debug-xgqwz\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.539360 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dccfc\" (UniqueName: \"kubernetes.io/projected/9104935c-53e2-4681-9e94-04a7354bba56-kube-api-access-dccfc\") pod \"crc-debug-xgqwz\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.640505 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9104935c-53e2-4681-9e94-04a7354bba56-host\") pod \"crc-debug-xgqwz\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.640622 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dccfc\" (UniqueName: \"kubernetes.io/projected/9104935c-53e2-4681-9e94-04a7354bba56-kube-api-access-dccfc\") pod \"crc-debug-xgqwz\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.640651 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9104935c-53e2-4681-9e94-04a7354bba56-host\") pod \"crc-debug-xgqwz\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.660352 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dccfc\" (UniqueName: \"kubernetes.io/projected/9104935c-53e2-4681-9e94-04a7354bba56-kube-api-access-dccfc\") pod \"crc-debug-xgqwz\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.699184 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.997587 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" event={"ID":"9104935c-53e2-4681-9e94-04a7354bba56","Type":"ContainerStarted","Data":"4088abec32f6ee4512dbc4550c0cae676573be1eafa5ff94653d083066842c32"} Feb 19 09:46:07 crc kubenswrapper[4675]: I0219 09:46:07.997915 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" event={"ID":"9104935c-53e2-4681-9e94-04a7354bba56","Type":"ContainerStarted","Data":"d56fef44299c12546c3e7846b3153e40486eab4fbe9dce0661fc9961f88e1967"} Feb 19 09:46:09 crc kubenswrapper[4675]: I0219 09:46:09.008599 4675 generic.go:334] "Generic (PLEG): container finished" podID="9104935c-53e2-4681-9e94-04a7354bba56" containerID="4088abec32f6ee4512dbc4550c0cae676573be1eafa5ff94653d083066842c32" exitCode=0 Feb 19 09:46:09 crc kubenswrapper[4675]: I0219 09:46:09.008648 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" event={"ID":"9104935c-53e2-4681-9e94-04a7354bba56","Type":"ContainerDied","Data":"4088abec32f6ee4512dbc4550c0cae676573be1eafa5ff94653d083066842c32"} Feb 19 09:46:09 crc kubenswrapper[4675]: I0219 09:46:09.487866 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-xgqwz"] Feb 19 09:46:09 crc kubenswrapper[4675]: I0219 09:46:09.500317 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-xgqwz"] Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.117395 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.290513 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dccfc\" (UniqueName: \"kubernetes.io/projected/9104935c-53e2-4681-9e94-04a7354bba56-kube-api-access-dccfc\") pod \"9104935c-53e2-4681-9e94-04a7354bba56\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.290633 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9104935c-53e2-4681-9e94-04a7354bba56-host\") pod \"9104935c-53e2-4681-9e94-04a7354bba56\" (UID: \"9104935c-53e2-4681-9e94-04a7354bba56\") " Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.291048 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9104935c-53e2-4681-9e94-04a7354bba56-host" (OuterVolumeSpecName: "host") pod "9104935c-53e2-4681-9e94-04a7354bba56" (UID: "9104935c-53e2-4681-9e94-04a7354bba56"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.291547 4675 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9104935c-53e2-4681-9e94-04a7354bba56-host\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.300685 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9104935c-53e2-4681-9e94-04a7354bba56-kube-api-access-dccfc" (OuterVolumeSpecName: "kube-api-access-dccfc") pod "9104935c-53e2-4681-9e94-04a7354bba56" (UID: "9104935c-53e2-4681-9e94-04a7354bba56"). InnerVolumeSpecName "kube-api-access-dccfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.393814 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dccfc\" (UniqueName: \"kubernetes.io/projected/9104935c-53e2-4681-9e94-04a7354bba56-kube-api-access-dccfc\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.709140 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-z58c2"] Feb 19 09:46:10 crc kubenswrapper[4675]: E0219 09:46:10.711443 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9104935c-53e2-4681-9e94-04a7354bba56" containerName="container-00" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.711573 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="9104935c-53e2-4681-9e94-04a7354bba56" containerName="container-00" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.711907 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="9104935c-53e2-4681-9e94-04a7354bba56" containerName="container-00" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.712894 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.806895 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/731c62ba-68c8-4016-bf2f-15a998aeee65-host\") pod \"crc-debug-z58c2\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.807312 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vn6j\" (UniqueName: \"kubernetes.io/projected/731c62ba-68c8-4016-bf2f-15a998aeee65-kube-api-access-9vn6j\") pod \"crc-debug-z58c2\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.908837 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/731c62ba-68c8-4016-bf2f-15a998aeee65-host\") pod \"crc-debug-z58c2\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.908921 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vn6j\" (UniqueName: \"kubernetes.io/projected/731c62ba-68c8-4016-bf2f-15a998aeee65-kube-api-access-9vn6j\") pod \"crc-debug-z58c2\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.908970 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/731c62ba-68c8-4016-bf2f-15a998aeee65-host\") pod \"crc-debug-z58c2\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:10 crc kubenswrapper[4675]: I0219 09:46:10.928406 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vn6j\" (UniqueName: \"kubernetes.io/projected/731c62ba-68c8-4016-bf2f-15a998aeee65-kube-api-access-9vn6j\") pod \"crc-debug-z58c2\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:11 crc kubenswrapper[4675]: I0219 09:46:11.032267 4675 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d56fef44299c12546c3e7846b3153e40486eab4fbe9dce0661fc9961f88e1967" Feb 19 09:46:11 crc kubenswrapper[4675]: I0219 09:46:11.032328 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-xgqwz" Feb 19 09:46:11 crc kubenswrapper[4675]: I0219 09:46:11.079159 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:11 crc kubenswrapper[4675]: W0219 09:46:11.107696 4675 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod731c62ba_68c8_4016_bf2f_15a998aeee65.slice/crio-8be07f42c810d9ad6ce960315823e4a56f996f5f17f74a630eedcee31469dff2 WatchSource:0}: Error finding container 8be07f42c810d9ad6ce960315823e4a56f996f5f17f74a630eedcee31469dff2: Status 404 returned error can't find the container with id 8be07f42c810d9ad6ce960315823e4a56f996f5f17f74a630eedcee31469dff2 Feb 19 09:46:11 crc kubenswrapper[4675]: I0219 09:46:11.113889 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9104935c-53e2-4681-9e94-04a7354bba56" path="/var/lib/kubelet/pods/9104935c-53e2-4681-9e94-04a7354bba56/volumes" Feb 19 09:46:12 crc kubenswrapper[4675]: I0219 09:46:12.042694 4675 generic.go:334] "Generic (PLEG): container finished" podID="731c62ba-68c8-4016-bf2f-15a998aeee65" containerID="6eed926ceede891140ccb12817c664a904c547e5024d4888d32e203cb0a7d852" exitCode=0 Feb 19 09:46:12 crc kubenswrapper[4675]: I0219 09:46:12.042784 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-z58c2" event={"ID":"731c62ba-68c8-4016-bf2f-15a998aeee65","Type":"ContainerDied","Data":"6eed926ceede891140ccb12817c664a904c547e5024d4888d32e203cb0a7d852"} Feb 19 09:46:12 crc kubenswrapper[4675]: I0219 09:46:12.042994 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/crc-debug-z58c2" event={"ID":"731c62ba-68c8-4016-bf2f-15a998aeee65","Type":"ContainerStarted","Data":"8be07f42c810d9ad6ce960315823e4a56f996f5f17f74a630eedcee31469dff2"} Feb 19 09:46:12 crc kubenswrapper[4675]: I0219 09:46:12.077664 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-z58c2"] Feb 19 09:46:12 crc kubenswrapper[4675]: I0219 09:46:12.085875 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b4gs/crc-debug-z58c2"] Feb 19 09:46:13 crc kubenswrapper[4675]: I0219 09:46:13.159925 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:13 crc kubenswrapper[4675]: I0219 09:46:13.252671 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vn6j\" (UniqueName: \"kubernetes.io/projected/731c62ba-68c8-4016-bf2f-15a998aeee65-kube-api-access-9vn6j\") pod \"731c62ba-68c8-4016-bf2f-15a998aeee65\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " Feb 19 09:46:13 crc kubenswrapper[4675]: I0219 09:46:13.252818 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/731c62ba-68c8-4016-bf2f-15a998aeee65-host\") pod \"731c62ba-68c8-4016-bf2f-15a998aeee65\" (UID: \"731c62ba-68c8-4016-bf2f-15a998aeee65\") " Feb 19 09:46:13 crc kubenswrapper[4675]: I0219 09:46:13.253096 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/731c62ba-68c8-4016-bf2f-15a998aeee65-host" (OuterVolumeSpecName: "host") pod "731c62ba-68c8-4016-bf2f-15a998aeee65" (UID: "731c62ba-68c8-4016-bf2f-15a998aeee65"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 19 09:46:13 crc kubenswrapper[4675]: I0219 09:46:13.254022 4675 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/731c62ba-68c8-4016-bf2f-15a998aeee65-host\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:13 crc kubenswrapper[4675]: I0219 09:46:13.259048 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/731c62ba-68c8-4016-bf2f-15a998aeee65-kube-api-access-9vn6j" (OuterVolumeSpecName: "kube-api-access-9vn6j") pod "731c62ba-68c8-4016-bf2f-15a998aeee65" (UID: "731c62ba-68c8-4016-bf2f-15a998aeee65"). InnerVolumeSpecName "kube-api-access-9vn6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:46:13 crc kubenswrapper[4675]: I0219 09:46:13.355882 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vn6j\" (UniqueName: \"kubernetes.io/projected/731c62ba-68c8-4016-bf2f-15a998aeee65-kube-api-access-9vn6j\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:14 crc kubenswrapper[4675]: I0219 09:46:14.060146 4675 scope.go:117] "RemoveContainer" containerID="6eed926ceede891140ccb12817c664a904c547e5024d4888d32e203cb0a7d852" Feb 19 09:46:14 crc kubenswrapper[4675]: I0219 09:46:14.060321 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/crc-debug-z58c2" Feb 19 09:46:15 crc kubenswrapper[4675]: I0219 09:46:15.121553 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="731c62ba-68c8-4016-bf2f-15a998aeee65" path="/var/lib/kubelet/pods/731c62ba-68c8-4016-bf2f-15a998aeee65/volumes" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.675232 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vglkv"] Feb 19 09:46:39 crc kubenswrapper[4675]: E0219 09:46:39.676157 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731c62ba-68c8-4016-bf2f-15a998aeee65" containerName="container-00" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.676171 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="731c62ba-68c8-4016-bf2f-15a998aeee65" containerName="container-00" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.676349 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="731c62ba-68c8-4016-bf2f-15a998aeee65" containerName="container-00" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.677715 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.693808 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vglkv"] Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.815093 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-utilities\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.815162 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-catalog-content\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.815239 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl4dw\" (UniqueName: \"kubernetes.io/projected/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-kube-api-access-fl4dw\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.916834 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl4dw\" (UniqueName: \"kubernetes.io/projected/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-kube-api-access-fl4dw\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.916947 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-utilities\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.917008 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-catalog-content\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.917689 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-catalog-content\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.918211 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-utilities\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:39 crc kubenswrapper[4675]: I0219 09:46:39.940438 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl4dw\" (UniqueName: \"kubernetes.io/projected/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-kube-api-access-fl4dw\") pod \"redhat-marketplace-vglkv\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:40 crc kubenswrapper[4675]: I0219 09:46:40.014064 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:40 crc kubenswrapper[4675]: I0219 09:46:40.558059 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vglkv"] Feb 19 09:46:40 crc kubenswrapper[4675]: I0219 09:46:40.933230 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b4d586cf4-9nm5g_c045dff4-aad5-44b8-a942-20a45f927859/barbican-api/0.log" Feb 19 09:46:40 crc kubenswrapper[4675]: I0219 09:46:40.977589 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b4d586cf4-9nm5g_c045dff4-aad5-44b8-a942-20a45f927859/barbican-api-log/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.159793 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-bcb894f54-l299f_8649bd53-643b-466d-a528-45d2aa799316/barbican-keystone-listener/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.207533 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-bcb894f54-l299f_8649bd53-643b-466d-a528-45d2aa799316/barbican-keystone-listener-log/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.363861 4675 generic.go:334] "Generic (PLEG): container finished" podID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerID="2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632" exitCode=0 Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.363906 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vglkv" event={"ID":"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1","Type":"ContainerDied","Data":"2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632"} Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.363930 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vglkv" event={"ID":"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1","Type":"ContainerStarted","Data":"12824ca7fc1e8785833765fff16025b55ccb166acf8a31b9f31cd655ad5867cd"} Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.495409 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7667dd7745-49ztz_5a4a2232-762f-4fc1-83fd-45d33b408e5b/barbican-worker/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.546555 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7667dd7745-49ztz_5a4a2232-762f-4fc1-83fd-45d33b408e5b/barbican-worker-log/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.675957 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-z2ppd_df5fab89-7323-474a-96b4-60dbd70e2c50/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.743132 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.743181 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.764249 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/ceilometer-central-agent/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.838820 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/ceilometer-notification-agent/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.904253 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/proxy-httpd/0.log" Feb 19 09:46:41 crc kubenswrapper[4675]: I0219 09:46:41.920548 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f19360d2-5c69-4ad4-b4d5-9098dbf5bd84/sg-core/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.098373 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b83eea9-3ec2-44bd-8498-5c5604e0eac7/cinder-api/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.129366 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b83eea9-3ec2-44bd-8498-5c5604e0eac7/cinder-api-log/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.295100 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e85d2864-3bca-456e-a6a2-b5f4bc385b15/cinder-scheduler/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.320562 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e85d2864-3bca-456e-a6a2-b5f4bc385b15/probe/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.373498 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vglkv" event={"ID":"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1","Type":"ContainerStarted","Data":"17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae"} Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.536404 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-8bd6k_8fa31ca4-5356-4a67-ac53-6bc4d9f0780f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.551064 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-j4f7z_380f6f00-e968-4289-8ea9-ed1c8ef7ff59/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.713491 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-8grnp_012e6c07-a778-4850-99e0-47346310663b/init/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.925234 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-8grnp_012e6c07-a778-4850-99e0-47346310663b/init/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.948434 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-8grnp_012e6c07-a778-4850-99e0-47346310663b/dnsmasq-dns/0.log" Feb 19 09:46:42 crc kubenswrapper[4675]: I0219 09:46:42.970163 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-fr44t_31e7b53a-04a2-4fd3-b18e-893da088a60f/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.119605 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0be92ca9-2a5c-4c33-b7e4-a440f157cd02/glance-httpd/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.189631 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0be92ca9-2a5c-4c33-b7e4-a440f157cd02/glance-log/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.295412 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91a2c1f8-322b-40e7-a577-ed775f9a1565/glance-httpd/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.349454 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_91a2c1f8-322b-40e7-a577-ed775f9a1565/glance-log/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.383472 4675 generic.go:334] "Generic (PLEG): container finished" podID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerID="17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae" exitCode=0 Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.383578 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vglkv" event={"ID":"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1","Type":"ContainerDied","Data":"17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae"} Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.620515 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64d55559f4-zwfgc_18d90d12-7724-438d-8cd4-9fbe70c8ee89/horizon/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.663038 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-5cts8_8dad42cd-cf8e-40d0-ab92-37c57c05d7fa/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.812720 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64d55559f4-zwfgc_18d90d12-7724-438d-8cd4-9fbe70c8ee89/horizon-log/0.log" Feb 19 09:46:43 crc kubenswrapper[4675]: I0219 09:46:43.915365 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g5kn7_55d678f4-667d-4d16-8f40-4faeab6d6c1a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:44 crc kubenswrapper[4675]: I0219 09:46:44.164569 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1e70d121-b4ec-4c18-be34-18d2eb7d529e/kube-state-metrics/0.log" Feb 19 09:46:44 crc kubenswrapper[4675]: I0219 09:46:44.205131 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-8594567c89-8xr7n_a6ae7ce1-9da5-49f4-8e32-0f1e0996080a/keystone-api/0.log" Feb 19 09:46:44 crc kubenswrapper[4675]: I0219 09:46:44.396007 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vglkv" event={"ID":"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1","Type":"ContainerStarted","Data":"e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7"} Feb 19 09:46:44 crc kubenswrapper[4675]: I0219 09:46:44.428371 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vglkv" podStartSLOduration=3.056809128 podStartE2EDuration="5.428354287s" podCreationTimestamp="2026-02-19 09:46:39 +0000 UTC" firstStartedPulling="2026-02-19 09:46:41.371976508 +0000 UTC m=+3782.999066776" lastFinishedPulling="2026-02-19 09:46:43.743521667 +0000 UTC m=+3785.370611935" observedRunningTime="2026-02-19 09:46:44.424624837 +0000 UTC m=+3786.051715105" watchObservedRunningTime="2026-02-19 09:46:44.428354287 +0000 UTC m=+3786.055444555" Feb 19 09:46:44 crc kubenswrapper[4675]: I0219 09:46:44.527798 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pknwk_12901b5c-afde-4017-86ba-363d50c78a9a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:44 crc kubenswrapper[4675]: I0219 09:46:44.843957 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57f4c8679f-grtkm_81ba53aa-e359-412c-abe4-b91662a9996a/neutron-api/0.log" Feb 19 09:46:44 crc kubenswrapper[4675]: I0219 09:46:44.895630 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57f4c8679f-grtkm_81ba53aa-e359-412c-abe4-b91662a9996a/neutron-httpd/0.log" Feb 19 09:46:45 crc kubenswrapper[4675]: I0219 09:46:45.309380 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xrlm9_185f89ae-500c-4bdf-81b1-e3627ab1d600/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:45 crc kubenswrapper[4675]: I0219 09:46:45.831688 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_657eb3f7-7b32-44db-b26b-c070b619cee7/nova-api-log/0.log" Feb 19 09:46:45 crc kubenswrapper[4675]: I0219 09:46:45.986079 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_408c0ffb-6a67-4351-ac43-8c5affe01d69/nova-cell0-conductor-conductor/0.log" Feb 19 09:46:46 crc kubenswrapper[4675]: I0219 09:46:46.151634 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_657eb3f7-7b32-44db-b26b-c070b619cee7/nova-api-api/0.log" Feb 19 09:46:46 crc kubenswrapper[4675]: I0219 09:46:46.612932 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_670df879-c26a-46f0-b385-5bb153734e18/nova-cell1-conductor-conductor/0.log" Feb 19 09:46:46 crc kubenswrapper[4675]: I0219 09:46:46.785954 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-r2x4h_9cee21fc-60b4-413c-b6b4-e75e813a8e9d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:46 crc kubenswrapper[4675]: I0219 09:46:46.808152 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_774eaf0e-fab7-4e49-b3de-b9e8ca212f66/nova-cell1-novncproxy-novncproxy/0.log" Feb 19 09:46:46 crc kubenswrapper[4675]: I0219 09:46:46.915876 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78796291-e993-4cea-90df-f44081e71bf9/nova-metadata-log/0.log" Feb 19 09:46:47 crc kubenswrapper[4675]: I0219 09:46:47.358156 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_de054e7d-f903-4b96-8394-7d12f6cfdb98/nova-scheduler-scheduler/0.log" Feb 19 09:46:47 crc kubenswrapper[4675]: I0219 09:46:47.371812 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3b301f22-5f49-490c-899a-f518d8da2174/mysql-bootstrap/0.log" Feb 19 09:46:47 crc kubenswrapper[4675]: I0219 09:46:47.518953 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3b301f22-5f49-490c-899a-f518d8da2174/mysql-bootstrap/0.log" Feb 19 09:46:47 crc kubenswrapper[4675]: I0219 09:46:47.624417 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3b301f22-5f49-490c-899a-f518d8da2174/galera/0.log" Feb 19 09:46:47 crc kubenswrapper[4675]: I0219 09:46:47.796892 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b91a4d30-10a1-4827-b4a7-2c620a2c2d9f/mysql-bootstrap/0.log" Feb 19 09:46:47 crc kubenswrapper[4675]: I0219 09:46:47.975290 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b91a4d30-10a1-4827-b4a7-2c620a2c2d9f/mysql-bootstrap/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.011697 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b91a4d30-10a1-4827-b4a7-2c620a2c2d9f/galera/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.183929 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2b0f27e5-af96-47cd-9550-dfa37033f584/openstackclient/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.270586 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lnndq_10a37557-1329-4995-86a8-4dc2e12f217c/openstack-network-exporter/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.327062 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78796291-e993-4cea-90df-f44081e71bf9/nova-metadata-metadata/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.550979 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovsdb-server-init/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.748880 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovsdb-server-init/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.779245 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovs-vswitchd/0.log" Feb 19 09:46:48 crc kubenswrapper[4675]: I0219 09:46:48.784306 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5dht_6dd9438b-4fad-4a66-8d08-df4816a19124/ovsdb-server/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.001859 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vlnhc_0b11fc83-5e56-43d3-a340-80686a5bdf35/ovn-controller/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.162137 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cghvt_5a7be99c-546f-4174-80ca-d72818c3ee43/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.470409 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d9d70128-744b-4075-ab12-f60f6c1838e7/openstack-network-exporter/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.637012 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d9d70128-744b-4075-ab12-f60f6c1838e7/ovn-northd/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.664245 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_482efc7c-8e44-4c25-903e-5c8b2cc58d0b/openstack-network-exporter/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.860622 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_482efc7c-8e44-4c25-903e-5c8b2cc58d0b/ovsdbserver-nb/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.886368 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2344b53c-5df6-4673-beff-7c3cdd28dc93/openstack-network-exporter/0.log" Feb 19 09:46:49 crc kubenswrapper[4675]: I0219 09:46:49.955197 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2344b53c-5df6-4673-beff-7c3cdd28dc93/ovsdbserver-sb/0.log" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.014122 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.014457 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.070702 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.257610 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7c56db696b-nwkrv_271c0509-6c25-44bf-8565-5d25488f67c7/placement-api/0.log" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.296703 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7c56db696b-nwkrv_271c0509-6c25-44bf-8565-5d25488f67c7/placement-log/0.log" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.434875 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4/setup-container/0.log" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.521962 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.577478 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vglkv"] Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.693055 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_60253071-539f-43bb-a335-b351dbac4b77/setup-container/0.log" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.698464 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4/rabbitmq/0.log" Feb 19 09:46:50 crc kubenswrapper[4675]: I0219 09:46:50.714628 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94c0b8d8-c9c1-42e6-8050-9ecd8ec903b4/setup-container/0.log" Feb 19 09:46:51 crc kubenswrapper[4675]: I0219 09:46:51.511430 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_60253071-539f-43bb-a335-b351dbac4b77/rabbitmq/0.log" Feb 19 09:46:51 crc kubenswrapper[4675]: I0219 09:46:51.515435 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_60253071-539f-43bb-a335-b351dbac4b77/setup-container/0.log" Feb 19 09:46:51 crc kubenswrapper[4675]: I0219 09:46:51.630410 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dhlxt_50eaff9d-656b-4e39-b450-bcc31570bc29/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:51 crc kubenswrapper[4675]: I0219 09:46:51.831396 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zkwd8_185defec-09d9-4992-bc76-fc2cc62a788e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.029182 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-crfpp_34f8029c-b96e-42c2-8005-ebc9988389b7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.080464 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-692m6_f3bba895-fd3d-44bf-9ea7-e631560bbc0a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.226542 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-w7zd2_36e6ad34-b4ec-4e70-895b-37b09f4508aa/ssh-known-hosts-edpm-deployment/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.425940 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-676ddcfdc-554zf_c331033a-dd3e-4871-8e7d-85081578aace/proxy-server/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.481956 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vglkv" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="registry-server" containerID="cri-o://e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7" gracePeriod=2 Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.555157 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-676ddcfdc-554zf_c331033a-dd3e-4871-8e7d-85081578aace/proxy-httpd/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.572107 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-s4x6m_dadf56d2-1dd1-4590-8fd7-0cc266af5377/swift-ring-rebalance/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.733784 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-auditor/0.log" Feb 19 09:46:52 crc kubenswrapper[4675]: I0219 09:46:52.747677 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-reaper/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.215109 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-replicator/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.228290 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-auditor/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.251502 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/account-server/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.347193 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-replicator/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.358093 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.477448 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl4dw\" (UniqueName: \"kubernetes.io/projected/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-kube-api-access-fl4dw\") pod \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.477505 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-utilities\") pod \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.477628 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-catalog-content\") pod \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\" (UID: \"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1\") " Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.487149 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-utilities" (OuterVolumeSpecName: "utilities") pod "09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" (UID: "09c31f4b-e5db-4696-ba78-e7c7a84cb0e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.519989 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-kube-api-access-fl4dw" (OuterVolumeSpecName: "kube-api-access-fl4dw") pod "09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" (UID: "09c31f4b-e5db-4696-ba78-e7c7a84cb0e1"). InnerVolumeSpecName "kube-api-access-fl4dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.542611 4675 generic.go:334] "Generic (PLEG): container finished" podID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerID="e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7" exitCode=0 Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.542670 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vglkv" event={"ID":"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1","Type":"ContainerDied","Data":"e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7"} Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.542697 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vglkv" event={"ID":"09c31f4b-e5db-4696-ba78-e7c7a84cb0e1","Type":"ContainerDied","Data":"12824ca7fc1e8785833765fff16025b55ccb166acf8a31b9f31cd655ad5867cd"} Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.542715 4675 scope.go:117] "RemoveContainer" containerID="e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.542906 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vglkv" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.556774 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" (UID: "09c31f4b-e5db-4696-ba78-e7c7a84cb0e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.575815 4675 scope.go:117] "RemoveContainer" containerID="17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.579413 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl4dw\" (UniqueName: \"kubernetes.io/projected/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-kube-api-access-fl4dw\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.579441 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.579454 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.616176 4675 scope.go:117] "RemoveContainer" containerID="2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.648715 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-updater/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.662138 4675 scope.go:117] "RemoveContainer" containerID="e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7" Feb 19 09:46:53 crc kubenswrapper[4675]: E0219 09:46:53.663241 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7\": container with ID starting with e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7 not found: ID does not exist" containerID="e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.663277 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7"} err="failed to get container status \"e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7\": rpc error: code = NotFound desc = could not find container \"e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7\": container with ID starting with e6a4869a477d2945b58941ea4738445528add819fb38bd795ea9a7f8915ed1d7 not found: ID does not exist" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.663298 4675 scope.go:117] "RemoveContainer" containerID="17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae" Feb 19 09:46:53 crc kubenswrapper[4675]: E0219 09:46:53.663574 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae\": container with ID starting with 17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae not found: ID does not exist" containerID="17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.663607 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae"} err="failed to get container status \"17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae\": rpc error: code = NotFound desc = could not find container \"17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae\": container with ID starting with 17f173eb730873dca5046cb72604920e523d41f50ebc42807334e9713e7f36ae not found: ID does not exist" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.663620 4675 scope.go:117] "RemoveContainer" containerID="2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632" Feb 19 09:46:53 crc kubenswrapper[4675]: E0219 09:46:53.664022 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632\": container with ID starting with 2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632 not found: ID does not exist" containerID="2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.664062 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632"} err="failed to get container status \"2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632\": rpc error: code = NotFound desc = could not find container \"2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632\": container with ID starting with 2e80aed18a54262b070b0e74a2f8bf26489b32fc807d9aaa191dd5df2c19e632 not found: ID does not exist" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.699077 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/container-server/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.711123 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-expirer/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.784212 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-auditor/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.936336 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vglkv"] Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.952279 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-server/0.log" Feb 19 09:46:53 crc kubenswrapper[4675]: I0219 09:46:53.967293 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vglkv"] Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.021942 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-replicator/0.log" Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.052657 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/object-updater/0.log" Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.082873 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/rsync/0.log" Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.166845 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_74a1ef5b-9639-489e-9dc7-2da01bffa6f2/swift-recon-cron/0.log" Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.332446 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-bkjvl_c1b6f896-16c6-40af-ae5f-4392213dacd3/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.397414 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5006db86-ae9d-4dbc-871f-4bc18f3a43f6/tempest-tests-tempest-tests-runner/0.log" Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.572700 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b255b445-e2d1-4d12-8a53-34e76f18dede/test-operator-logs-container/0.log" Feb 19 09:46:54 crc kubenswrapper[4675]: I0219 09:46:54.610680 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qrnwj_af6f5807-47f8-4e48-b8c0-b04eeb4bb160/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 19 09:46:55 crc kubenswrapper[4675]: I0219 09:46:55.113754 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" path="/var/lib/kubelet/pods/09c31f4b-e5db-4696-ba78-e7c7a84cb0e1/volumes" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.412991 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nmfsl"] Feb 19 09:46:56 crc kubenswrapper[4675]: E0219 09:46:56.413754 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="registry-server" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.413770 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="registry-server" Feb 19 09:46:56 crc kubenswrapper[4675]: E0219 09:46:56.413790 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="extract-content" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.413796 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="extract-content" Feb 19 09:46:56 crc kubenswrapper[4675]: E0219 09:46:56.413816 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="extract-utilities" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.413822 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="extract-utilities" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.414010 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="09c31f4b-e5db-4696-ba78-e7c7a84cb0e1" containerName="registry-server" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.415336 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.425035 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nmfsl"] Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.526983 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-utilities\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.527047 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkc52\" (UniqueName: \"kubernetes.io/projected/15ecb8b5-941f-47d2-ba5a-401682f7540a-kube-api-access-kkc52\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.527244 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-catalog-content\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.628824 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-utilities\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.628871 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkc52\" (UniqueName: \"kubernetes.io/projected/15ecb8b5-941f-47d2-ba5a-401682f7540a-kube-api-access-kkc52\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.628987 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-catalog-content\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.629410 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-catalog-content\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.629864 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-utilities\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.656621 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkc52\" (UniqueName: \"kubernetes.io/projected/15ecb8b5-941f-47d2-ba5a-401682f7540a-kube-api-access-kkc52\") pod \"certified-operators-nmfsl\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:56 crc kubenswrapper[4675]: I0219 09:46:56.758103 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:46:57 crc kubenswrapper[4675]: I0219 09:46:57.344920 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nmfsl"] Feb 19 09:46:57 crc kubenswrapper[4675]: I0219 09:46:57.576852 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerStarted","Data":"aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f"} Feb 19 09:46:57 crc kubenswrapper[4675]: I0219 09:46:57.576893 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerStarted","Data":"3224ca873e3f9bd2f7e44013d122cd903bde6d9a517a845a3e0f2a6f6d191f90"} Feb 19 09:46:58 crc kubenswrapper[4675]: I0219 09:46:58.590846 4675 generic.go:334] "Generic (PLEG): container finished" podID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerID="aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f" exitCode=0 Feb 19 09:46:58 crc kubenswrapper[4675]: I0219 09:46:58.591338 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerDied","Data":"aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f"} Feb 19 09:46:59 crc kubenswrapper[4675]: I0219 09:46:59.602605 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerStarted","Data":"a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48"} Feb 19 09:47:00 crc kubenswrapper[4675]: I0219 09:47:00.611388 4675 generic.go:334] "Generic (PLEG): container finished" podID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerID="a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48" exitCode=0 Feb 19 09:47:00 crc kubenswrapper[4675]: I0219 09:47:00.611739 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerDied","Data":"a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48"} Feb 19 09:47:01 crc kubenswrapper[4675]: I0219 09:47:01.625088 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerStarted","Data":"360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506"} Feb 19 09:47:01 crc kubenswrapper[4675]: I0219 09:47:01.677232 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nmfsl" podStartSLOduration=3.022115953 podStartE2EDuration="5.677213201s" podCreationTimestamp="2026-02-19 09:46:56 +0000 UTC" firstStartedPulling="2026-02-19 09:46:58.596634917 +0000 UTC m=+3800.223725185" lastFinishedPulling="2026-02-19 09:47:01.251732165 +0000 UTC m=+3802.878822433" observedRunningTime="2026-02-19 09:47:01.649695187 +0000 UTC m=+3803.276785455" watchObservedRunningTime="2026-02-19 09:47:01.677213201 +0000 UTC m=+3803.304303459" Feb 19 09:47:02 crc kubenswrapper[4675]: I0219 09:47:02.271946 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4f51af02-6352-4462-b00f-a2feafe19f23/memcached/0.log" Feb 19 09:47:06 crc kubenswrapper[4675]: I0219 09:47:06.758700 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:47:06 crc kubenswrapper[4675]: I0219 09:47:06.759231 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:47:06 crc kubenswrapper[4675]: I0219 09:47:06.812953 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:47:07 crc kubenswrapper[4675]: I0219 09:47:07.730368 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:47:07 crc kubenswrapper[4675]: I0219 09:47:07.827520 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nmfsl"] Feb 19 09:47:09 crc kubenswrapper[4675]: I0219 09:47:09.688761 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nmfsl" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="registry-server" containerID="cri-o://360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506" gracePeriod=2 Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.164576 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.278468 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-catalog-content\") pod \"15ecb8b5-941f-47d2-ba5a-401682f7540a\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.278627 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkc52\" (UniqueName: \"kubernetes.io/projected/15ecb8b5-941f-47d2-ba5a-401682f7540a-kube-api-access-kkc52\") pod \"15ecb8b5-941f-47d2-ba5a-401682f7540a\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.278694 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-utilities\") pod \"15ecb8b5-941f-47d2-ba5a-401682f7540a\" (UID: \"15ecb8b5-941f-47d2-ba5a-401682f7540a\") " Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.280447 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-utilities" (OuterVolumeSpecName: "utilities") pod "15ecb8b5-941f-47d2-ba5a-401682f7540a" (UID: "15ecb8b5-941f-47d2-ba5a-401682f7540a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.287773 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15ecb8b5-941f-47d2-ba5a-401682f7540a-kube-api-access-kkc52" (OuterVolumeSpecName: "kube-api-access-kkc52") pod "15ecb8b5-941f-47d2-ba5a-401682f7540a" (UID: "15ecb8b5-941f-47d2-ba5a-401682f7540a"). InnerVolumeSpecName "kube-api-access-kkc52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.340783 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15ecb8b5-941f-47d2-ba5a-401682f7540a" (UID: "15ecb8b5-941f-47d2-ba5a-401682f7540a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.381744 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.381788 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkc52\" (UniqueName: \"kubernetes.io/projected/15ecb8b5-941f-47d2-ba5a-401682f7540a-kube-api-access-kkc52\") on node \"crc\" DevicePath \"\"" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.381804 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ecb8b5-941f-47d2-ba5a-401682f7540a-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.699319 4675 generic.go:334] "Generic (PLEG): container finished" podID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerID="360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506" exitCode=0 Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.699364 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerDied","Data":"360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506"} Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.699393 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmfsl" event={"ID":"15ecb8b5-941f-47d2-ba5a-401682f7540a","Type":"ContainerDied","Data":"3224ca873e3f9bd2f7e44013d122cd903bde6d9a517a845a3e0f2a6f6d191f90"} Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.699414 4675 scope.go:117] "RemoveContainer" containerID="360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.699540 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmfsl" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.744804 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nmfsl"] Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.747482 4675 scope.go:117] "RemoveContainer" containerID="a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.752993 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nmfsl"] Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.770475 4675 scope.go:117] "RemoveContainer" containerID="aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.813727 4675 scope.go:117] "RemoveContainer" containerID="360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506" Feb 19 09:47:10 crc kubenswrapper[4675]: E0219 09:47:10.814128 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506\": container with ID starting with 360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506 not found: ID does not exist" containerID="360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.814158 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506"} err="failed to get container status \"360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506\": rpc error: code = NotFound desc = could not find container \"360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506\": container with ID starting with 360272be20a35788954068cd3c3325771f617b9e928ac4274c8727da75bfb506 not found: ID does not exist" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.814179 4675 scope.go:117] "RemoveContainer" containerID="a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48" Feb 19 09:47:10 crc kubenswrapper[4675]: E0219 09:47:10.814475 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48\": container with ID starting with a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48 not found: ID does not exist" containerID="a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.814518 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48"} err="failed to get container status \"a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48\": rpc error: code = NotFound desc = could not find container \"a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48\": container with ID starting with a97d5dc468ca148ae6b3c1527c3759005518dfd248d0aba4d8e488d8263f9b48 not found: ID does not exist" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.814555 4675 scope.go:117] "RemoveContainer" containerID="aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f" Feb 19 09:47:10 crc kubenswrapper[4675]: E0219 09:47:10.814866 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f\": container with ID starting with aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f not found: ID does not exist" containerID="aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f" Feb 19 09:47:10 crc kubenswrapper[4675]: I0219 09:47:10.814889 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f"} err="failed to get container status \"aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f\": rpc error: code = NotFound desc = could not find container \"aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f\": container with ID starting with aacc7f3acb8f4388a324754887c7f26d720bbfa0bcdd0e87c4cd35bb3eefe39f not found: ID does not exist" Feb 19 09:47:11 crc kubenswrapper[4675]: I0219 09:47:11.113545 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" path="/var/lib/kubelet/pods/15ecb8b5-941f-47d2-ba5a-401682f7540a/volumes" Feb 19 09:47:11 crc kubenswrapper[4675]: I0219 09:47:11.743577 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:47:11 crc kubenswrapper[4675]: I0219 09:47:11.743687 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:47:20 crc kubenswrapper[4675]: I0219 09:47:20.609893 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/util/0.log" Feb 19 09:47:20 crc kubenswrapper[4675]: I0219 09:47:20.704288 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/util/0.log" Feb 19 09:47:20 crc kubenswrapper[4675]: I0219 09:47:20.776367 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/pull/0.log" Feb 19 09:47:20 crc kubenswrapper[4675]: I0219 09:47:20.809100 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/pull/0.log" Feb 19 09:47:21 crc kubenswrapper[4675]: I0219 09:47:21.010055 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/pull/0.log" Feb 19 09:47:21 crc kubenswrapper[4675]: I0219 09:47:21.027046 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/extract/0.log" Feb 19 09:47:21 crc kubenswrapper[4675]: I0219 09:47:21.028672 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40ad46de3d09c6788a0a61961f484c0fd3241032aa57f1a594be8a7d1b22ntl_f2fb95eb-0086-4762-96c1-fe6956a852e2/util/0.log" Feb 19 09:47:21 crc kubenswrapper[4675]: I0219 09:47:21.452216 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-glh49_7f7fd1c2-d277-402d-a762-0ebfddc48226/manager/0.log" Feb 19 09:47:21 crc kubenswrapper[4675]: I0219 09:47:21.802946 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-v7lcl_b0de8493-3f97-4e5e-a28d-5d45f545d645/manager/0.log" Feb 19 09:47:22 crc kubenswrapper[4675]: I0219 09:47:22.027629 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-bwwwp_53318e37-b20a-46eb-8f12-decb087ade91/manager/0.log" Feb 19 09:47:22 crc kubenswrapper[4675]: I0219 09:47:22.239590 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-9zn82_2473d2e8-7c4d-41f5-bd2a-95823fe059f1/manager/0.log" Feb 19 09:47:22 crc kubenswrapper[4675]: I0219 09:47:22.588297 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-c775s_f79e0726-e5c1-4536-b06b-40bab3849bb0/manager/0.log" Feb 19 09:47:22 crc kubenswrapper[4675]: I0219 09:47:22.750202 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-z2wq2_78209850-b313-4190-a3a4-674da1afaaba/manager/0.log" Feb 19 09:47:22 crc kubenswrapper[4675]: I0219 09:47:22.921661 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-cbbkp_a5bfbb7f-0d4a-49d3-bf77-b168ccf33061/manager/0.log" Feb 19 09:47:23 crc kubenswrapper[4675]: I0219 09:47:23.011490 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-jhgw4_4a22a41e-3e7a-4c2a-93ce-d37bbaf434b6/manager/0.log" Feb 19 09:47:23 crc kubenswrapper[4675]: I0219 09:47:23.152761 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-kqz2s_63fcb9d8-c313-4d0e-9f53-44c3ad272d1d/manager/0.log" Feb 19 09:47:23 crc kubenswrapper[4675]: I0219 09:47:23.355399 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-c25kr_61877abc-2f6b-40a5-aa19-b8adf5402cc6/manager/0.log" Feb 19 09:47:23 crc kubenswrapper[4675]: I0219 09:47:23.539537 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-8k6xs_5c9124b1-4ca0-481f-8474-ea45252bda76/manager/0.log" Feb 19 09:47:23 crc kubenswrapper[4675]: I0219 09:47:23.690224 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-ddhjt_4e163d5d-3666-4899-acd7-21040c1cc573/manager/0.log" Feb 19 09:47:24 crc kubenswrapper[4675]: I0219 09:47:24.039125 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cjkhd8_769e3e7a-a9ea-4d9c-b60b-8643e0c275e1/manager/0.log" Feb 19 09:47:24 crc kubenswrapper[4675]: I0219 09:47:24.402971 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-767d85d5f6-7tx82_f7d62133-059d-4695-829f-ed7b9888e323/operator/0.log" Feb 19 09:47:24 crc kubenswrapper[4675]: I0219 09:47:24.716045 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-h9qft_8aa0b598-0a09-4439-9e9d-c125fb5542db/registry-server/0.log" Feb 19 09:47:24 crc kubenswrapper[4675]: I0219 09:47:24.998515 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-r8bkh_dcd57b86-afb5-4beb-ae73-830ce75a562f/manager/0.log" Feb 19 09:47:25 crc kubenswrapper[4675]: I0219 09:47:25.181684 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-dr2nj_2f4e8661-38a2-44bc-90ae-266f058758a9/manager/0.log" Feb 19 09:47:25 crc kubenswrapper[4675]: I0219 09:47:25.230815 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-ff4m9_4fad9fc9-cb88-4aff-89bb-0a2fd609f48a/manager/0.log" Feb 19 09:47:25 crc kubenswrapper[4675]: I0219 09:47:25.434930 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-sm2fn_b9dc6b16-af6b-470e-a9a7-d59245e8a22e/operator/0.log" Feb 19 09:47:25 crc kubenswrapper[4675]: I0219 09:47:25.461923 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-jz6fw_53989876-5869-4e41-a9c3-c99cea8347bd/manager/0.log" Feb 19 09:47:25 crc kubenswrapper[4675]: I0219 09:47:25.704695 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-s25g8_b1bbb05f-c65b-4ea1-ab19-34b888f49369/manager/0.log" Feb 19 09:47:25 crc kubenswrapper[4675]: I0219 09:47:25.848403 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-w8w5t_70a13bde-19ab-4c7c-af9a-15f1fe96d4e1/manager/0.log" Feb 19 09:47:25 crc kubenswrapper[4675]: I0219 09:47:25.956567 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-x6glz_1a383e65-12c2-43bf-b6b0-f61f20706d7a/manager/0.log" Feb 19 09:47:26 crc kubenswrapper[4675]: I0219 09:47:26.270744 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7d685cfdd5-zzm7g_a4559dd6-5152-40d8-94af-f2fff6c90e84/manager/0.log" Feb 19 09:47:27 crc kubenswrapper[4675]: I0219 09:47:27.891225 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-bj7kj_2f6f91c7-8cd8-4a33-bf0f-cb8cf1a2e282/manager/0.log" Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.743691 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.744223 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.744294 4675 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.745322 4675 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a"} pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.745417 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" containerID="cri-o://6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" gracePeriod=600 Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.953367 4675 generic.go:334] "Generic (PLEG): container finished" podID="5e91469d-12b7-434c-991d-633287712a69" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" exitCode=0 Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.953741 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerDied","Data":"6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a"} Feb 19 09:47:41 crc kubenswrapper[4675]: I0219 09:47:41.953785 4675 scope.go:117] "RemoveContainer" containerID="e250573cac06229729a4b58db31b629a73c1df55fa477fcc7c737ad0061e5267" Feb 19 09:47:41 crc kubenswrapper[4675]: E0219 09:47:41.964218 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:47:42 crc kubenswrapper[4675]: I0219 09:47:42.963927 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:47:42 crc kubenswrapper[4675]: E0219 09:47:42.964575 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:47:45 crc kubenswrapper[4675]: I0219 09:47:45.229056 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-wf8zz_bbed8b9a-ef34-424b-81e0-457a38384ba0/control-plane-machine-set-operator/0.log" Feb 19 09:47:45 crc kubenswrapper[4675]: I0219 09:47:45.416706 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6tbj_7c39c460-3de5-4b4a-9c38-76ad1de9cbce/kube-rbac-proxy/0.log" Feb 19 09:47:45 crc kubenswrapper[4675]: I0219 09:47:45.444687 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6tbj_7c39c460-3de5-4b4a-9c38-76ad1de9cbce/machine-api-operator/0.log" Feb 19 09:47:56 crc kubenswrapper[4675]: I0219 09:47:56.103330 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:47:56 crc kubenswrapper[4675]: E0219 09:47:56.104489 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:47:58 crc kubenswrapper[4675]: I0219 09:47:58.309393 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-mfc4d_2138c669-3d8b-45c6-9c6d-dfc5d9ef0d2e/cert-manager-controller/0.log" Feb 19 09:47:58 crc kubenswrapper[4675]: I0219 09:47:58.496954 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-5sxn8_bf9ebddc-98fd-4515-8354-f1d7ae9fda4d/cert-manager-webhook/0.log" Feb 19 09:47:58 crc kubenswrapper[4675]: I0219 09:47:58.503951 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-z9x24_68aab37f-51c0-4132-844e-bc47b4856f16/cert-manager-cainjector/0.log" Feb 19 09:48:10 crc kubenswrapper[4675]: I0219 09:48:10.820940 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-f84lw_e6d09f36-de74-42bb-8b27-3ac3039ee87f/nmstate-console-plugin/0.log" Feb 19 09:48:11 crc kubenswrapper[4675]: I0219 09:48:11.013778 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-k28cx_d1523822-5fa6-446b-9cd1-3442536238f2/nmstate-handler/0.log" Feb 19 09:48:11 crc kubenswrapper[4675]: I0219 09:48:11.106497 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:48:11 crc kubenswrapper[4675]: E0219 09:48:11.106830 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:48:11 crc kubenswrapper[4675]: I0219 09:48:11.107974 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-92cqz_8a261a06-8401-488e-8b98-d885f1b694cf/nmstate-metrics/0.log" Feb 19 09:48:11 crc kubenswrapper[4675]: I0219 09:48:11.138284 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-92cqz_8a261a06-8401-488e-8b98-d885f1b694cf/kube-rbac-proxy/0.log" Feb 19 09:48:11 crc kubenswrapper[4675]: I0219 09:48:11.219602 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-mmtc8_a6fb0b89-c8be-4b62-a92b-23197720e978/nmstate-operator/0.log" Feb 19 09:48:11 crc kubenswrapper[4675]: I0219 09:48:11.336765 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-chcgg_3be88498-1577-460a-9431-4a4bc66eb217/nmstate-webhook/0.log" Feb 19 09:48:26 crc kubenswrapper[4675]: I0219 09:48:26.103363 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:48:26 crc kubenswrapper[4675]: E0219 09:48:26.104980 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:48:37 crc kubenswrapper[4675]: I0219 09:48:37.898134 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-j4tln_47ca0e9d-940d-4952-9780-74e58570d98e/kube-rbac-proxy/0.log" Feb 19 09:48:37 crc kubenswrapper[4675]: I0219 09:48:37.950299 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-j4tln_47ca0e9d-940d-4952-9780-74e58570d98e/controller/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.211597 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.382570 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.421352 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.424820 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.463119 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.652204 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.657518 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.657676 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.689378 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.862039 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-reloader/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.871323 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-metrics/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.920817 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/controller/0.log" Feb 19 09:48:38 crc kubenswrapper[4675]: I0219 09:48:38.944229 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/cp-frr-files/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.060286 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/frr-metrics/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.145327 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/kube-rbac-proxy/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.172994 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/kube-rbac-proxy-frr/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.301515 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/reloader/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.455280 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-fjr4t_ee8232ef-f035-4ef4-b867-98508be2d0e7/frr-k8s-webhook-server/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.539728 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-698bb84f65-v6wc9_8e1476c5-559f-4da9-9092-ef6b08253b6a/manager/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.742998 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-85dcc8c87f-62fjw_d7d966d2-c2fa-411d-8474-ab26f3cbea46/webhook-server/0.log" Feb 19 09:48:39 crc kubenswrapper[4675]: I0219 09:48:39.869440 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bc8vc_75ab6240-327c-4ac4-94f2-3d4c44e1a664/kube-rbac-proxy/0.log" Feb 19 09:48:40 crc kubenswrapper[4675]: I0219 09:48:40.103355 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:48:40 crc kubenswrapper[4675]: E0219 09:48:40.103630 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:48:40 crc kubenswrapper[4675]: I0219 09:48:40.432098 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bc8vc_75ab6240-327c-4ac4-94f2-3d4c44e1a664/speaker/0.log" Feb 19 09:48:40 crc kubenswrapper[4675]: I0219 09:48:40.668474 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fj4cv_e0bff342-a422-4fe7-b4cb-ca04e88bf905/frr/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.306918 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/util/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.499068 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/util/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.507524 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/pull/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.551667 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/pull/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.748611 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/util/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.772536 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/pull/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.772784 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s96r8_7388a06e-7e03-4aec-8326-6bcfe1cc194a/extract/0.log" Feb 19 09:48:53 crc kubenswrapper[4675]: I0219 09:48:53.923007 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-utilities/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.079405 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-content/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.109254 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:48:54 crc kubenswrapper[4675]: E0219 09:48:54.109721 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.126177 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-utilities/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.143705 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-content/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.310519 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-utilities/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.323741 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/extract-content/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.535034 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-utilities/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.780093 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mvrwn_50ddb4cd-7719-4f7a-a71f-a8bd832236e0/registry-server/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.846307 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-utilities/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.853629 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-content/0.log" Feb 19 09:48:54 crc kubenswrapper[4675]: I0219 09:48:54.860904 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-content/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.021613 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-utilities/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.036607 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/extract-content/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.213001 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/util/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.561031 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gmkvv_4adf2b0e-ea04-47fb-8c06-e8f2902d1257/registry-server/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.843759 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/util/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.844682 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/pull/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.893171 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/pull/0.log" Feb 19 09:48:55 crc kubenswrapper[4675]: I0219 09:48:55.976691 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/util/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.028861 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/extract/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.029429 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecasdwqk_c07488ba-ffa0-479e-8c03-bb41ae1b24f3/pull/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.177192 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-c4g5q_147346d6-fe78-404a-964e-4ee12c505b82/marketplace-operator/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.194008 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-utilities/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.385982 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-utilities/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.421073 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-content/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.443994 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-content/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.599086 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-content/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.618071 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/extract-utilities/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.739663 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fg4v6_3d9580d3-8d3f-4730-af25-6c57eead5415/registry-server/0.log" Feb 19 09:48:56 crc kubenswrapper[4675]: I0219 09:48:56.933955 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-utilities/0.log" Feb 19 09:48:57 crc kubenswrapper[4675]: I0219 09:48:57.411575 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-utilities/0.log" Feb 19 09:48:57 crc kubenswrapper[4675]: I0219 09:48:57.509366 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-content/0.log" Feb 19 09:48:57 crc kubenswrapper[4675]: I0219 09:48:57.533309 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-content/0.log" Feb 19 09:48:57 crc kubenswrapper[4675]: I0219 09:48:57.684612 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-content/0.log" Feb 19 09:48:57 crc kubenswrapper[4675]: I0219 09:48:57.743295 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/extract-utilities/0.log" Feb 19 09:48:58 crc kubenswrapper[4675]: I0219 09:48:58.149750 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8xtr4_b2b9392d-292a-40aa-9271-805bde5fd1d6/registry-server/0.log" Feb 19 09:49:09 crc kubenswrapper[4675]: I0219 09:49:09.103733 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:49:09 crc kubenswrapper[4675]: E0219 09:49:09.104474 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:49:22 crc kubenswrapper[4675]: I0219 09:49:22.103283 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:49:22 crc kubenswrapper[4675]: E0219 09:49:22.104129 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:49:36 crc kubenswrapper[4675]: I0219 09:49:36.103179 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:49:36 crc kubenswrapper[4675]: E0219 09:49:36.103899 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:49:48 crc kubenswrapper[4675]: I0219 09:49:48.104323 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:49:48 crc kubenswrapper[4675]: E0219 09:49:48.105414 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:49:59 crc kubenswrapper[4675]: I0219 09:49:59.117797 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:49:59 crc kubenswrapper[4675]: E0219 09:49:59.119980 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:50:14 crc kubenswrapper[4675]: I0219 09:50:14.104693 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:50:14 crc kubenswrapper[4675]: E0219 09:50:14.105468 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:50:29 crc kubenswrapper[4675]: I0219 09:50:29.109439 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:50:29 crc kubenswrapper[4675]: E0219 09:50:29.110174 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:50:41 crc kubenswrapper[4675]: I0219 09:50:41.103675 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:50:41 crc kubenswrapper[4675]: E0219 09:50:41.105904 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:50:44 crc kubenswrapper[4675]: I0219 09:50:44.555232 4675 generic.go:334] "Generic (PLEG): container finished" podID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerID="597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc" exitCode=0 Feb 19 09:50:44 crc kubenswrapper[4675]: I0219 09:50:44.555300 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" event={"ID":"fab9926e-edf2-4c5a-bed3-40d59f293458","Type":"ContainerDied","Data":"597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc"} Feb 19 09:50:44 crc kubenswrapper[4675]: I0219 09:50:44.556459 4675 scope.go:117] "RemoveContainer" containerID="597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc" Feb 19 09:50:44 crc kubenswrapper[4675]: I0219 09:50:44.857663 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7b4gs_must-gather-rxbnt_fab9926e-edf2-4c5a-bed3-40d59f293458/gather/0.log" Feb 19 09:50:54 crc kubenswrapper[4675]: I0219 09:50:54.103242 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:50:54 crc kubenswrapper[4675]: E0219 09:50:54.103896 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:50:55 crc kubenswrapper[4675]: I0219 09:50:55.628412 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b4gs/must-gather-rxbnt"] Feb 19 09:50:55 crc kubenswrapper[4675]: I0219 09:50:55.629078 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerName="copy" containerID="cri-o://7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b" gracePeriod=2 Feb 19 09:50:55 crc kubenswrapper[4675]: I0219 09:50:55.638680 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b4gs/must-gather-rxbnt"] Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.032271 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7b4gs_must-gather-rxbnt_fab9926e-edf2-4c5a-bed3-40d59f293458/copy/0.log" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.032821 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.227787 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fab9926e-edf2-4c5a-bed3-40d59f293458-must-gather-output\") pod \"fab9926e-edf2-4c5a-bed3-40d59f293458\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.227917 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nh8s\" (UniqueName: \"kubernetes.io/projected/fab9926e-edf2-4c5a-bed3-40d59f293458-kube-api-access-9nh8s\") pod \"fab9926e-edf2-4c5a-bed3-40d59f293458\" (UID: \"fab9926e-edf2-4c5a-bed3-40d59f293458\") " Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.234570 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fab9926e-edf2-4c5a-bed3-40d59f293458-kube-api-access-9nh8s" (OuterVolumeSpecName: "kube-api-access-9nh8s") pod "fab9926e-edf2-4c5a-bed3-40d59f293458" (UID: "fab9926e-edf2-4c5a-bed3-40d59f293458"). InnerVolumeSpecName "kube-api-access-9nh8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.331463 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nh8s\" (UniqueName: \"kubernetes.io/projected/fab9926e-edf2-4c5a-bed3-40d59f293458-kube-api-access-9nh8s\") on node \"crc\" DevicePath \"\"" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.413202 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fab9926e-edf2-4c5a-bed3-40d59f293458-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fab9926e-edf2-4c5a-bed3-40d59f293458" (UID: "fab9926e-edf2-4c5a-bed3-40d59f293458"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.433377 4675 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fab9926e-edf2-4c5a-bed3-40d59f293458-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.674919 4675 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7b4gs_must-gather-rxbnt_fab9926e-edf2-4c5a-bed3-40d59f293458/copy/0.log" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.675261 4675 generic.go:334] "Generic (PLEG): container finished" podID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerID="7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b" exitCode=143 Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.675313 4675 scope.go:117] "RemoveContainer" containerID="7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.675334 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b4gs/must-gather-rxbnt" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.696175 4675 scope.go:117] "RemoveContainer" containerID="597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.753966 4675 scope.go:117] "RemoveContainer" containerID="7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b" Feb 19 09:50:56 crc kubenswrapper[4675]: E0219 09:50:56.754742 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b\": container with ID starting with 7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b not found: ID does not exist" containerID="7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.754816 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b"} err="failed to get container status \"7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b\": rpc error: code = NotFound desc = could not find container \"7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b\": container with ID starting with 7127ffe691634ed1f6dfb27dc57d5b9342c4660f22fdf78acfb27b4a5b5d0c1b not found: ID does not exist" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.754857 4675 scope.go:117] "RemoveContainer" containerID="597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc" Feb 19 09:50:56 crc kubenswrapper[4675]: E0219 09:50:56.755436 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc\": container with ID starting with 597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc not found: ID does not exist" containerID="597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc" Feb 19 09:50:56 crc kubenswrapper[4675]: I0219 09:50:56.755466 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc"} err="failed to get container status \"597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc\": rpc error: code = NotFound desc = could not find container \"597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc\": container with ID starting with 597e46b5c032179a99515f997ace45e8de38eac017b5ab05df1769786541a2fc not found: ID does not exist" Feb 19 09:50:57 crc kubenswrapper[4675]: I0219 09:50:57.115507 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" path="/var/lib/kubelet/pods/fab9926e-edf2-4c5a-bed3-40d59f293458/volumes" Feb 19 09:51:06 crc kubenswrapper[4675]: I0219 09:51:06.103357 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:51:06 crc kubenswrapper[4675]: E0219 09:51:06.104102 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:51:21 crc kubenswrapper[4675]: I0219 09:51:21.103376 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:51:21 crc kubenswrapper[4675]: E0219 09:51:21.104113 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:51:32 crc kubenswrapper[4675]: I0219 09:51:32.103731 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:51:32 crc kubenswrapper[4675]: E0219 09:51:32.104724 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:51:47 crc kubenswrapper[4675]: I0219 09:51:47.103792 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:51:47 crc kubenswrapper[4675]: E0219 09:51:47.104606 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:51:59 crc kubenswrapper[4675]: I0219 09:51:59.556281 4675 scope.go:117] "RemoveContainer" containerID="4005f0753a3bf37f95ee6c7bf9b917b9cfddf700a95228e9733838d2762d75c0" Feb 19 09:52:02 crc kubenswrapper[4675]: I0219 09:52:02.103810 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:52:02 crc kubenswrapper[4675]: E0219 09:52:02.104676 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:52:15 crc kubenswrapper[4675]: I0219 09:52:15.103466 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:52:15 crc kubenswrapper[4675]: E0219 09:52:15.104117 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:52:30 crc kubenswrapper[4675]: I0219 09:52:30.103922 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:52:30 crc kubenswrapper[4675]: E0219 09:52:30.106099 4675 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wqmkn_openshift-machine-config-operator(5e91469d-12b7-434c-991d-633287712a69)\"" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" Feb 19 09:52:42 crc kubenswrapper[4675]: I0219 09:52:42.103610 4675 scope.go:117] "RemoveContainer" containerID="6473001ce270a9c2f0f26e064700407a633e48347ee62083273d8de659c1d39a" Feb 19 09:52:42 crc kubenswrapper[4675]: I0219 09:52:42.586272 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" event={"ID":"5e91469d-12b7-434c-991d-633287712a69","Type":"ContainerStarted","Data":"bca62d90fd5ffb84a69c12a8dc19502af2152ba5d7f47aa30a0656a115b7eaa3"} Feb 19 09:52:59 crc kubenswrapper[4675]: I0219 09:52:59.605715 4675 scope.go:117] "RemoveContainer" containerID="4088abec32f6ee4512dbc4550c0cae676573be1eafa5ff94653d083066842c32" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.848606 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tqv4t"] Feb 19 09:53:02 crc kubenswrapper[4675]: E0219 09:53:02.849577 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerName="gather" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849594 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerName="gather" Feb 19 09:53:02 crc kubenswrapper[4675]: E0219 09:53:02.849607 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="registry-server" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849615 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="registry-server" Feb 19 09:53:02 crc kubenswrapper[4675]: E0219 09:53:02.849629 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="extract-utilities" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849660 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="extract-utilities" Feb 19 09:53:02 crc kubenswrapper[4675]: E0219 09:53:02.849684 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="extract-content" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849692 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="extract-content" Feb 19 09:53:02 crc kubenswrapper[4675]: E0219 09:53:02.849712 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerName="copy" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849720 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerName="copy" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849923 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerName="gather" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849942 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="fab9926e-edf2-4c5a-bed3-40d59f293458" containerName="copy" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.849959 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ecb8b5-941f-47d2-ba5a-401682f7540a" containerName="registry-server" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.851611 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.871253 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqv4t"] Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.890960 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-utilities\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.891022 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdhsw\" (UniqueName: \"kubernetes.io/projected/270a9041-9541-4641-a6cf-9b03a66cff38-kube-api-access-tdhsw\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.891052 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-catalog-content\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.994153 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-utilities\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.994262 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdhsw\" (UniqueName: \"kubernetes.io/projected/270a9041-9541-4641-a6cf-9b03a66cff38-kube-api-access-tdhsw\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.994312 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-catalog-content\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.994995 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-utilities\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:02 crc kubenswrapper[4675]: I0219 09:53:02.995061 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-catalog-content\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:03 crc kubenswrapper[4675]: I0219 09:53:03.014334 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdhsw\" (UniqueName: \"kubernetes.io/projected/270a9041-9541-4641-a6cf-9b03a66cff38-kube-api-access-tdhsw\") pod \"redhat-operators-tqv4t\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:03 crc kubenswrapper[4675]: I0219 09:53:03.174408 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:03 crc kubenswrapper[4675]: I0219 09:53:03.779582 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqv4t"] Feb 19 09:53:04 crc kubenswrapper[4675]: I0219 09:53:04.790437 4675 generic.go:334] "Generic (PLEG): container finished" podID="270a9041-9541-4641-a6cf-9b03a66cff38" containerID="e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe" exitCode=0 Feb 19 09:53:04 crc kubenswrapper[4675]: I0219 09:53:04.790507 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqv4t" event={"ID":"270a9041-9541-4641-a6cf-9b03a66cff38","Type":"ContainerDied","Data":"e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe"} Feb 19 09:53:04 crc kubenswrapper[4675]: I0219 09:53:04.790950 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqv4t" event={"ID":"270a9041-9541-4641-a6cf-9b03a66cff38","Type":"ContainerStarted","Data":"3bc1f4e00d8f413a38ebbd4cd3f2597755003a66cceb02c44a83c2f5432992ee"} Feb 19 09:53:04 crc kubenswrapper[4675]: I0219 09:53:04.793598 4675 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 19 09:53:05 crc kubenswrapper[4675]: I0219 09:53:05.802194 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqv4t" event={"ID":"270a9041-9541-4641-a6cf-9b03a66cff38","Type":"ContainerStarted","Data":"0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af"} Feb 19 09:53:06 crc kubenswrapper[4675]: I0219 09:53:06.812488 4675 generic.go:334] "Generic (PLEG): container finished" podID="270a9041-9541-4641-a6cf-9b03a66cff38" containerID="0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af" exitCode=0 Feb 19 09:53:06 crc kubenswrapper[4675]: I0219 09:53:06.812542 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqv4t" event={"ID":"270a9041-9541-4641-a6cf-9b03a66cff38","Type":"ContainerDied","Data":"0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af"} Feb 19 09:53:08 crc kubenswrapper[4675]: I0219 09:53:08.841902 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqv4t" event={"ID":"270a9041-9541-4641-a6cf-9b03a66cff38","Type":"ContainerStarted","Data":"2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e"} Feb 19 09:53:08 crc kubenswrapper[4675]: I0219 09:53:08.868316 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tqv4t" podStartSLOduration=4.36384841 podStartE2EDuration="6.868295082s" podCreationTimestamp="2026-02-19 09:53:02 +0000 UTC" firstStartedPulling="2026-02-19 09:53:04.793387495 +0000 UTC m=+4166.420477763" lastFinishedPulling="2026-02-19 09:53:07.297834167 +0000 UTC m=+4168.924924435" observedRunningTime="2026-02-19 09:53:08.858965933 +0000 UTC m=+4170.486056201" watchObservedRunningTime="2026-02-19 09:53:08.868295082 +0000 UTC m=+4170.495385350" Feb 19 09:53:13 crc kubenswrapper[4675]: I0219 09:53:13.174880 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:13 crc kubenswrapper[4675]: I0219 09:53:13.175569 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:14 crc kubenswrapper[4675]: I0219 09:53:14.219190 4675 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tqv4t" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="registry-server" probeResult="failure" output=< Feb 19 09:53:14 crc kubenswrapper[4675]: timeout: failed to connect service ":50051" within 1s Feb 19 09:53:14 crc kubenswrapper[4675]: > Feb 19 09:53:23 crc kubenswrapper[4675]: I0219 09:53:23.354606 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:23 crc kubenswrapper[4675]: I0219 09:53:23.398453 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:23 crc kubenswrapper[4675]: I0219 09:53:23.588537 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqv4t"] Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.313655 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tqv4t" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="registry-server" containerID="cri-o://2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e" gracePeriod=2 Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.753444 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.860424 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-utilities\") pod \"270a9041-9541-4641-a6cf-9b03a66cff38\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.860492 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-catalog-content\") pod \"270a9041-9541-4641-a6cf-9b03a66cff38\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.860606 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdhsw\" (UniqueName: \"kubernetes.io/projected/270a9041-9541-4641-a6cf-9b03a66cff38-kube-api-access-tdhsw\") pod \"270a9041-9541-4641-a6cf-9b03a66cff38\" (UID: \"270a9041-9541-4641-a6cf-9b03a66cff38\") " Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.862314 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-utilities" (OuterVolumeSpecName: "utilities") pod "270a9041-9541-4641-a6cf-9b03a66cff38" (UID: "270a9041-9541-4641-a6cf-9b03a66cff38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.866087 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270a9041-9541-4641-a6cf-9b03a66cff38-kube-api-access-tdhsw" (OuterVolumeSpecName: "kube-api-access-tdhsw") pod "270a9041-9541-4641-a6cf-9b03a66cff38" (UID: "270a9041-9541-4641-a6cf-9b03a66cff38"). InnerVolumeSpecName "kube-api-access-tdhsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.963056 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdhsw\" (UniqueName: \"kubernetes.io/projected/270a9041-9541-4641-a6cf-9b03a66cff38-kube-api-access-tdhsw\") on node \"crc\" DevicePath \"\"" Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.963087 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:53:25 crc kubenswrapper[4675]: I0219 09:53:25.992903 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "270a9041-9541-4641-a6cf-9b03a66cff38" (UID: "270a9041-9541-4641-a6cf-9b03a66cff38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.064762 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270a9041-9541-4641-a6cf-9b03a66cff38-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.327292 4675 generic.go:334] "Generic (PLEG): container finished" podID="270a9041-9541-4641-a6cf-9b03a66cff38" containerID="2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e" exitCode=0 Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.327492 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqv4t" event={"ID":"270a9041-9541-4641-a6cf-9b03a66cff38","Type":"ContainerDied","Data":"2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e"} Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.327671 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqv4t" event={"ID":"270a9041-9541-4641-a6cf-9b03a66cff38","Type":"ContainerDied","Data":"3bc1f4e00d8f413a38ebbd4cd3f2597755003a66cceb02c44a83c2f5432992ee"} Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.327694 4675 scope.go:117] "RemoveContainer" containerID="2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.327609 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqv4t" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.352243 4675 scope.go:117] "RemoveContainer" containerID="0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.364381 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqv4t"] Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.373410 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tqv4t"] Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.385326 4675 scope.go:117] "RemoveContainer" containerID="e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.427397 4675 scope.go:117] "RemoveContainer" containerID="2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e" Feb 19 09:53:26 crc kubenswrapper[4675]: E0219 09:53:26.427904 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e\": container with ID starting with 2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e not found: ID does not exist" containerID="2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.427938 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e"} err="failed to get container status \"2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e\": rpc error: code = NotFound desc = could not find container \"2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e\": container with ID starting with 2a5fb9ee4a4db358be452abff274420a5299211c3674e262aed81a1442bfe25e not found: ID does not exist" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.427977 4675 scope.go:117] "RemoveContainer" containerID="0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af" Feb 19 09:53:26 crc kubenswrapper[4675]: E0219 09:53:26.429296 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af\": container with ID starting with 0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af not found: ID does not exist" containerID="0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.429329 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af"} err="failed to get container status \"0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af\": rpc error: code = NotFound desc = could not find container \"0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af\": container with ID starting with 0cc9b3b8b15cc80f43f01317b1c21489d2ea99882cf1e5f5527821b838c3e1af not found: ID does not exist" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.429346 4675 scope.go:117] "RemoveContainer" containerID="e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe" Feb 19 09:53:26 crc kubenswrapper[4675]: E0219 09:53:26.429626 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe\": container with ID starting with e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe not found: ID does not exist" containerID="e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe" Feb 19 09:53:26 crc kubenswrapper[4675]: I0219 09:53:26.429675 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe"} err="failed to get container status \"e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe\": rpc error: code = NotFound desc = could not find container \"e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe\": container with ID starting with e5cc51ee6403bb8e967dc5bcf029a5dd89f57d0b8780c58e621bacee26ec30fe not found: ID does not exist" Feb 19 09:53:27 crc kubenswrapper[4675]: I0219 09:53:27.112962 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" path="/var/lib/kubelet/pods/270a9041-9541-4641-a6cf-9b03a66cff38/volumes" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.160412 4675 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d98bx"] Feb 19 09:54:26 crc kubenswrapper[4675]: E0219 09:54:26.165172 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="extract-utilities" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.165205 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="extract-utilities" Feb 19 09:54:26 crc kubenswrapper[4675]: E0219 09:54:26.165237 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="registry-server" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.165245 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="registry-server" Feb 19 09:54:26 crc kubenswrapper[4675]: E0219 09:54:26.165277 4675 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="extract-content" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.165286 4675 state_mem.go:107] "Deleted CPUSet assignment" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="extract-content" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.165593 4675 memory_manager.go:354] "RemoveStaleState removing state" podUID="270a9041-9541-4641-a6cf-9b03a66cff38" containerName="registry-server" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.167235 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.180107 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d98bx"] Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.240511 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-utilities\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.240954 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-catalog-content\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.241061 4675 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zztmv\" (UniqueName: \"kubernetes.io/projected/f8583369-4538-4b73-ab68-2311f77ce3eb-kube-api-access-zztmv\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.343105 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-utilities\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.343234 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-catalog-content\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.343266 4675 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zztmv\" (UniqueName: \"kubernetes.io/projected/f8583369-4538-4b73-ab68-2311f77ce3eb-kube-api-access-zztmv\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.343519 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-utilities\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.343818 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-catalog-content\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.374632 4675 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zztmv\" (UniqueName: \"kubernetes.io/projected/f8583369-4538-4b73-ab68-2311f77ce3eb-kube-api-access-zztmv\") pod \"community-operators-d98bx\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.487608 4675 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:26 crc kubenswrapper[4675]: I0219 09:54:26.996668 4675 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d98bx"] Feb 19 09:54:27 crc kubenswrapper[4675]: I0219 09:54:27.907009 4675 generic.go:334] "Generic (PLEG): container finished" podID="f8583369-4538-4b73-ab68-2311f77ce3eb" containerID="f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8" exitCode=0 Feb 19 09:54:27 crc kubenswrapper[4675]: I0219 09:54:27.907103 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d98bx" event={"ID":"f8583369-4538-4b73-ab68-2311f77ce3eb","Type":"ContainerDied","Data":"f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8"} Feb 19 09:54:27 crc kubenswrapper[4675]: I0219 09:54:27.907277 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d98bx" event={"ID":"f8583369-4538-4b73-ab68-2311f77ce3eb","Type":"ContainerStarted","Data":"dc83a347c8a6bc04a2d78951927ee18d5c8feacf4f9cf0b6a899004d61c1053c"} Feb 19 09:54:28 crc kubenswrapper[4675]: I0219 09:54:28.917795 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d98bx" event={"ID":"f8583369-4538-4b73-ab68-2311f77ce3eb","Type":"ContainerStarted","Data":"a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a"} Feb 19 09:54:29 crc kubenswrapper[4675]: I0219 09:54:29.928353 4675 generic.go:334] "Generic (PLEG): container finished" podID="f8583369-4538-4b73-ab68-2311f77ce3eb" containerID="a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a" exitCode=0 Feb 19 09:54:29 crc kubenswrapper[4675]: I0219 09:54:29.928394 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d98bx" event={"ID":"f8583369-4538-4b73-ab68-2311f77ce3eb","Type":"ContainerDied","Data":"a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a"} Feb 19 09:54:30 crc kubenswrapper[4675]: I0219 09:54:30.940395 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d98bx" event={"ID":"f8583369-4538-4b73-ab68-2311f77ce3eb","Type":"ContainerStarted","Data":"6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149"} Feb 19 09:54:30 crc kubenswrapper[4675]: I0219 09:54:30.961953 4675 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d98bx" podStartSLOduration=2.566392506 podStartE2EDuration="4.961929735s" podCreationTimestamp="2026-02-19 09:54:26 +0000 UTC" firstStartedPulling="2026-02-19 09:54:27.908860445 +0000 UTC m=+4249.535950713" lastFinishedPulling="2026-02-19 09:54:30.304397674 +0000 UTC m=+4251.931487942" observedRunningTime="2026-02-19 09:54:30.957553168 +0000 UTC m=+4252.584643436" watchObservedRunningTime="2026-02-19 09:54:30.961929735 +0000 UTC m=+4252.589020003" Feb 19 09:54:36 crc kubenswrapper[4675]: I0219 09:54:36.488513 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:36 crc kubenswrapper[4675]: I0219 09:54:36.489037 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:36 crc kubenswrapper[4675]: I0219 09:54:36.542614 4675 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:37 crc kubenswrapper[4675]: I0219 09:54:37.049915 4675 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:37 crc kubenswrapper[4675]: I0219 09:54:37.124577 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d98bx"] Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.000427 4675 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d98bx" podUID="f8583369-4538-4b73-ab68-2311f77ce3eb" containerName="registry-server" containerID="cri-o://6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149" gracePeriod=2 Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.451935 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.534361 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-utilities\") pod \"f8583369-4538-4b73-ab68-2311f77ce3eb\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.534437 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zztmv\" (UniqueName: \"kubernetes.io/projected/f8583369-4538-4b73-ab68-2311f77ce3eb-kube-api-access-zztmv\") pod \"f8583369-4538-4b73-ab68-2311f77ce3eb\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.534529 4675 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-catalog-content\") pod \"f8583369-4538-4b73-ab68-2311f77ce3eb\" (UID: \"f8583369-4538-4b73-ab68-2311f77ce3eb\") " Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.541839 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-utilities" (OuterVolumeSpecName: "utilities") pod "f8583369-4538-4b73-ab68-2311f77ce3eb" (UID: "f8583369-4538-4b73-ab68-2311f77ce3eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.546840 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8583369-4538-4b73-ab68-2311f77ce3eb-kube-api-access-zztmv" (OuterVolumeSpecName: "kube-api-access-zztmv") pod "f8583369-4538-4b73-ab68-2311f77ce3eb" (UID: "f8583369-4538-4b73-ab68-2311f77ce3eb"). InnerVolumeSpecName "kube-api-access-zztmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.583398 4675 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8583369-4538-4b73-ab68-2311f77ce3eb" (UID: "f8583369-4538-4b73-ab68-2311f77ce3eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.636344 4675 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.636372 4675 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8583369-4538-4b73-ab68-2311f77ce3eb-utilities\") on node \"crc\" DevicePath \"\"" Feb 19 09:54:39 crc kubenswrapper[4675]: I0219 09:54:39.636383 4675 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zztmv\" (UniqueName: \"kubernetes.io/projected/f8583369-4538-4b73-ab68-2311f77ce3eb-kube-api-access-zztmv\") on node \"crc\" DevicePath \"\"" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.010112 4675 generic.go:334] "Generic (PLEG): container finished" podID="f8583369-4538-4b73-ab68-2311f77ce3eb" containerID="6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149" exitCode=0 Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.010164 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d98bx" event={"ID":"f8583369-4538-4b73-ab68-2311f77ce3eb","Type":"ContainerDied","Data":"6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149"} Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.010191 4675 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d98bx" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.010205 4675 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d98bx" event={"ID":"f8583369-4538-4b73-ab68-2311f77ce3eb","Type":"ContainerDied","Data":"dc83a347c8a6bc04a2d78951927ee18d5c8feacf4f9cf0b6a899004d61c1053c"} Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.010227 4675 scope.go:117] "RemoveContainer" containerID="6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.040513 4675 scope.go:117] "RemoveContainer" containerID="a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.045295 4675 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d98bx"] Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.060499 4675 scope.go:117] "RemoveContainer" containerID="f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.061493 4675 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d98bx"] Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.104631 4675 scope.go:117] "RemoveContainer" containerID="6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149" Feb 19 09:54:40 crc kubenswrapper[4675]: E0219 09:54:40.105158 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149\": container with ID starting with 6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149 not found: ID does not exist" containerID="6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.105187 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149"} err="failed to get container status \"6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149\": rpc error: code = NotFound desc = could not find container \"6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149\": container with ID starting with 6ac1d9b3b3c57548d799d38721c1bd663fe1da0febfac237aa7a477888b0c149 not found: ID does not exist" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.105205 4675 scope.go:117] "RemoveContainer" containerID="a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a" Feb 19 09:54:40 crc kubenswrapper[4675]: E0219 09:54:40.105601 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a\": container with ID starting with a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a not found: ID does not exist" containerID="a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.105624 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a"} err="failed to get container status \"a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a\": rpc error: code = NotFound desc = could not find container \"a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a\": container with ID starting with a885efd9a6498166535e3b5b895b38e850b6e7bafca6c80efbcbe20e7037547a not found: ID does not exist" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.105651 4675 scope.go:117] "RemoveContainer" containerID="f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8" Feb 19 09:54:40 crc kubenswrapper[4675]: E0219 09:54:40.105910 4675 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8\": container with ID starting with f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8 not found: ID does not exist" containerID="f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8" Feb 19 09:54:40 crc kubenswrapper[4675]: I0219 09:54:40.105929 4675 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8"} err="failed to get container status \"f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8\": rpc error: code = NotFound desc = could not find container \"f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8\": container with ID starting with f95b94203caac828c00cb622f1d64a12f234fde3ae2c10d7d19522ad8a4524f8 not found: ID does not exist" Feb 19 09:54:41 crc kubenswrapper[4675]: I0219 09:54:41.114841 4675 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8583369-4538-4b73-ab68-2311f77ce3eb" path="/var/lib/kubelet/pods/f8583369-4538-4b73-ab68-2311f77ce3eb/volumes" Feb 19 09:55:11 crc kubenswrapper[4675]: I0219 09:55:11.744107 4675 patch_prober.go:28] interesting pod/machine-config-daemon-wqmkn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 19 09:55:11 crc kubenswrapper[4675]: I0219 09:55:11.744702 4675 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wqmkn" podUID="5e91469d-12b7-434c-991d-633287712a69" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515145557014024454 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015145557015017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015145546172016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015145546172015467 5ustar corecore